Posted: April 25th, 2025
How do I manage organizational performance?
7 to 10 pages
Evaluation and Program Planning 61 (2017) 125–127
May Britt Bjerke*, Ralph Renger
University of North Dakota, School of Medicine & Health Sciences, Center for Rural Health Evaluation, 250 Centennial Dr. Stop 8138, Grand Forks, ND 58202-
8138, United States
A R T I C L E I N F O
Article history:
Received 17 October 2016
Received in revised form 16 December 2016
Accepted 19 December 2016
Available online 23 December 2016
Keywords:
Objective development
Mainstreaming
Evaluation guidance
SMART objectives
A B S T R A C T
This article challenges the conventional wisdom in mainstream evaluation regarding the process for
developing specific, measurable, attainable, relevant, and time-bound (SMART) objectives. The article
notes several advantages of mainstreaming the SMART method including program capacity building and
being able to independently monitor progress toward process and outcome objectives. It is argued the
one size fits all approach for writing SMART objectives is misleading. The context in which the evaluation
is conducted is a key deciding factor in how and when the SMART criteria should be applied. Without an
appreciation of the evaluation context, mainstream users may be developing objectives that are far from
smart. A case example is presented demonstrating a situation where a stepwise, rather than
simultaneous application of the SMART criteria was necessary. Learning from this case, recommen-
dations are forwarded for adjusting how SMART criteria should be presented in mainstream evaluation
manuals/guides.
© 2016 Elsevier Ltd. All rights reserved.
Contents lists available at ScienceDirect
Evaluation and Program Planning
journal homepage: www.elsevier .com/ locate /eva lprogplan
1. Introduction
Doran (1981) first introduced the specific, measurable, assign-
able, realistic, and time-related (SMART) method for writing
effective management goals. Today, the SMART method in
management is commonly stated as the standard for developing
effective, measurable goals and objectives (Bowles, Cunningham,
De La Rosa, & Picano, 2007; Conzemius & O’Neill, 2011; Frey &
Osterloh, 2001; Gettman, 2008; Hessel, Cortese, & De Croon, 2011;
Hofman & Hofman, 2011; Jung, 2007; MacLeod, 2012; Lawlor,
2012; Linstrom, 2006; Pearson, 2012; Piskurich, 2015; van der Grift
et al., 2013; Wade, 2009).
Although developed within management, the SMART method is
also widely cited within the program planning/evaluation litera-
ture (Chen, 2015; Gudda, 2011; Isell, 2014; Knowlton & Philips,
2013; Mathison, 2005; Patton, 2011; Sharma & Petosa, 2012; Smith,
2010). Moreover, program planning/evaluation guides provided by
the Centers for Disease Control and Prevention, the United Way,
The W.K. Kellogg Foundation and the United States Department of
Education, include the recommendation of using SMART criteria
when creating program goals and objectives (Bryan, DiMartino, &
Center for Secondary School Redesign, 2010; Centers for Disease
* Corresponding author.
E-mail addresses: maybritt.bjerke@med.und.edu (M.B. Bjerke),
ralph.renger@med.und.edu (R. Renger).
http://dx.doi.org/10.1016/j.evalprogplan.2016.12.009
0149-7189/© 2016 Elsevier Ltd. All rights reserved.
Control and Prevention, 2013; Harris and Harvard Family Research
Project, 2011; W.K. Kellogg Foundation, 2004). The proliferation of
the SMART method in evaluation and non-profit organization
guidance supports the contention that SMART is now a main-
stream method for developing program goals and objectives.
The benefit of mainstreaming is a greater number of programs,
especially those with limited resources, are able to apply
evaluation fundamentals to monitor and make program improve-
ments (Picciotto, 2002; Preskill & Boyle, 2008; Sanders, 2002). This
increased evaluation capacity reduces the need for costly external
evaluation consultants (Cousins, Goh, Elliott, Aubry, & Gilbert,
2014; Picciotto, 2002). It also enables more programs to meet
funders’ evaluation requirements (Stevenson, Florin, Mills, &
Andrade, 2002).
However, as is the case with attempting to mainstream any
evaluation method, there are many potential unintended con-
sequences (Grudens-Schuck, 2003; Merton & Sztompka, 1996;
Picciotto, 2002; Renger, 2006; Williams & Hawkes, 2003). First,
many mainstream program evaluation guides present the SMART
criteria without an explanation as to why or how they should be
applied. Thus, users may “blindly” following the recipe-like
method to develop SMART objectives without fully understanding
the underlying reasons for applying each SMART criterion. Second,
when following a recipe-like formula writing SMART objectives
may become nothing more than a grantsmanship exercise; a
necessary box needing to be checked to fulfill a sponsor’s request
for proposal requirements. Hummelbrunner (2010) expressed
http://crossmark.crossref.org/dialog/?doi=10.1016/j.evalprogplan.2016.12.009&domain=pdf
mailto:maybritt.bjerke@med.und.edu
mailto:ralph.renger@med.und.edu
mailto:ralph.renger@med.und.edu
http://dx.doi.org/10.1016/j.evalprogplan.2016.12.009
http://dx.doi.org/10.1016/j.evalprogplan.2016.12.009
http://www.sciencedirect.com/science/journal/01497189
www.elsevier.com/locate/evalprogplan
126 M.B. Bjerke, R. Renger / Evaluation and Program Planning 61 (2017) 125–127
similar concerns that laypersons following mainstream guidance
often do so as a justification rather than a planning exercise.
Finally, and specifically to mainstreaming SMART objectives,
program evaluation guides suggest SMART objectives be written in
a single step. On the surface, this may seem reasonable and
harmless. However, it is the authors’ contention there are some
instances where attempting to satisfy all the SMART criteria in a
single step is unrealistic and/or unwise. This method may produce
a mechanical approach to program evaluation objective writing.
The following case example demonstrates a situation where a
stepwise approach, rather than a simultaneous application of the
SMART criteria, was necessary to write meaningful program
objectives.
2. Case example
The authors’ need to elaborate on an alternative method for
developing SMART objectives arose while working on a self-
assessment tool for a cardiac ready communities (CRC) program in
a rural Midwest state (Center for Rural Health, 2016). The goal of a
CRC (also known as Heart Safe communities) program is to
increase survival rates from out-of-hospital cardiac arrest (OHCA)
through several community strategies targeting the five links in
the American Heart Association’s “chain of survival” (American
Heart Association, 2015; Heart Safe Communities, n.d.): compris-
ing 1) recognition of cardiac arrest and activation of the emergency
response system, 2) immediate cardiopulmonary resuscitation
(CPR), 3) rapid defibrillation, 4) basic and advanced emergency
medical services, and 5) advanced life support and post-cardiac
arrest care.
The CRC strategies are designed to address one or more of the
survival links and may include: community leadership involve-
ment, community awareness campaigns, CPR training, public
access to automated external defibrillators (AEDs), emergency
medical dispatching, resuscitation protocols for emergency medi-
cal services (EMS) and hospital services, and community evalua-
tion (Heart Safe Communities, n.d.; Montana Cardiac Ready
Communities, 2015; North Dakota Department of Health, 2016).
The state established the success of each strategy by providing
targets needing to be met within the three-year program time-
frame to receive official recognition as a cardiac ready community.
Within the state, there were numerous CRC’s needing evalua-
tion assistance. Further, given the numerous strategies encom-
passed in a single CRC initiative, it was not feasible to provide each
community with the external evaluation resources needed to track
their individual progress in meeting the set program targets.
Therefore, it was decided the best evaluation strategy was to
empower participating communities to conduct their own CRC
evaluation.
The evaluation strategy consisted of providing participating
communities with evaluation tools and technical assistance to
enable ongoing self-assessment (Center for Rural Health, 2016).
The self-assessment tool included guidance on how to create
SMART objectives for each CRC program activity so communities
could track progress toward the program targets (Center for Rural
Health, 2016). Specifically, the initial draft of the self-assessment
guide described how and why to write specific, measurable,
achievable, relevant, and timely objectives in line with the SMART
criteria suggested by Chen (2015).
The process of writing the guide forced evaluators to a deeper
level of thinking as to how the SMART criteria would be applied.
Explaining how to make the objectives specific, measurable, and
relevant was relatively straightforward. For example, one program
strategy related to the link of early CPR is community level CPR
training. To meet the specificity criterion the community needed to
detail what was meant by the terms “trained” and “population”.
For instance, “trained” could mean the population is at a minimum
trained in hands-only CPR within the last two years and
“population” could be defined as all community members aged
10 and above. To meet the measurable criterion the number of
community members trained in CPR could simply be tracked via
CPR course attendance sheets. The objective was relevant because
of the research evidence linking change in this essential link in the
chain to improved OHCA survival rates (American Heart Associa-
tion, 2015).
However, challenges arose when attempting to explain how to
apply the achievable and timely criteria. An achievable objective is
one that can be reasonably met with existing resources (Chen,
2015). Thus, whether an objective is achievable depends on having
the needed resources to move from the baseline to the desired goal.
For example, assume the baseline revealed 20% of the community
was CPR trained, but the goal was to have 25% trained. If both
budgets and training resources such as instructors, training
materials, and manikins were limited, then the 25% target might
be an attainable target. Alternatively, if the community had access
to ample resources, then perhaps a higher target such as 35% being
CPR trained might be achievable.
To meet the timely criterion requires objectives to include a
reference date for completion. Doing so, according to Chen (2015),
stimulates effectiveness. Although the state imposed a three-year
time frame, some objectives needed to be completed sooner than
others. However, without a baseline assessment establishing a
reasonable timeframe was challenging. For example, it is reason-
able to posit the community would achieve the 25% target sooner if
20% of the population was already trained in CPR, as compared to if
the baseline was closer to 10%. If the former was the case, the
SMART objective could state that by year 2 of the three-year
program the community will increase from 20% to 25% the share of
community members aged 10 and above trained in at least hands-
only CPR.
These challenges made it clear the self-assessment tool needed
to be modified so the CRCs (i) initially apply the criteria specific,
measurable, and relevant to their objectives, (ii) then gather
baseline data (because measurable has been defined), and (iii)
finally add to the objective quality by applying the achievable and
timely criteria. This held true for all strategies for each link in the
survival chain. The revised self-assessment tool provided more
detailed guidance aiding the communities in applying each SMART
criterion (Center for Rural Health, 2016). For example, adding
descriptions/definitions of all program strategies helped commu-
nities in adding specificity to their SMART objectives. Further,
adding available community resources provided assistance in
developing achievable objectives with realistic timeframes. SMART
objectives are more likely when a formal planning, implementa-
tion, and evaluation process like the Antecedent Target Measure-
ment (ATM) approach are followed with fidelity (Renger &
Titcomb, 2002). The revisions to the self-assessment tool better
assisted stakeholders in writing program objectives that met all
the SMART criteria. However, as reviewers of our work rightly note
additional guidance could be provided for each SMART criterion.
Currently, the self-assessment tool is being revisited to see where
decision rules could be added. It is a continuous improvement
process and the authors are getting smarter around their SMART
objectives guidance.
3. Conclusion
The case presentation demonstrates a uniform, one step SMART
approach may not always result in smart objectives. In our
example, the absence of baseline information did not allow for the
writing of achievable and timely program objectives. Thus, there
was a need for a stepwise approach to creating SMART objectives.
M.B. Bjerke, R. Renger / Evaluation and Program Planning 61 (2017) 125–127 127
The stakeholders first wrote specific, measurable, and relevant
objectives; then gathered baseline data. Once the baseline data
were collected the achievable and timely criteria could be applied.
While two steps were needed in this context, it is possible to
imagine situations where perhaps additional steps are needed
before satisfying all the SMART criteria.
3.1. Lessons learned
Learning from this, mainstreaming SMART objectives must be
done with some degree of caution and account for users who do
not fully understand the underlying reasons for applying each
SMART criterion. It is important future mainstream evaluation
manuals/guides delineate between different contexts in their
SMART goal/objective guidance as stakeholders may need to do
some homework before being able to satisfy all the SMART criteria.
Otherwise, programs may end up with SMART objectives that are
not so smart after all.
Acknowledgements
The authors would like to thank Dr. Carlos Rodriguez, Kim
Dickman, Skyler Ienuso, Makenzie McPherson, Allyssa Schlosser
and Eric Souvannasacd for their insights and feedback.
This work was supported by the Leona M. and Harry B. Helmsley
Charitable Trust.
References
American Heart Association (2015). Highlights of the 2015 American Heart Association
guidelines update for CPR and ECG. Retrieved 15 December 2016 from https://
eccguidelines.heart.org/wp-content/uploads/2015/10/2015-AHA-Guidelines-
Highlights-English .
Bowles, S., Cunningham, C. J., De La Rosa, G. M., & Picano, J. (2007). Coaching leaders
in middle and executive management: Goals, performance, buy-in. Leadership &
Organization Development Journal, 28(5), 388–408.
Bryan, W., DiMartino, J., & Center for Secondary School Redesign (2010). Writing
goals and objectives. A guide for grantees of the smaller learning communities
program. United States Department of Education, Smaller Learning
Communities Program. Academy for Educational Development Retrieved 15
December 2016 from https://www2.ed.gov/programs/slcp/slc-wgandobj-book-
f .
Center for Rural Health (2016). Cardiac ready community implementation and
evaluation guidelines. School of Medicine and Health Sciences, University of
North Dakota Unpublished document.
Centers for Disease Control, & Prevention (2013). Evaluation guide. Writing SMART
objectives. Division for Heart Disease and Stroke Prevention. Sate Heart Disease
and Stroke Prevention Program Retrieved 15 December 2016 from http://www.
cdc.gov/dhdsp/programs/spha/evaluation_guides/docs/smart_objectives .
Chen, H. T. (2015). Practical program evaluation: Theory-driven evaluation and the
integrated evaluation perspective, 2nd ed. Thousand Oaks, CA: Sage.
Conzemius, A., & O’Neill, J. (2011). The power of SMART goals: Using goals to improve
student learning. Solution Tree Press.
Cousins, J. B., Goh, S. C., Elliott, C., Aubry, T., & Gilbert, N. (2014). Government and
voluntary sector differences in organizational capacity to do and use evaluation.
Evaluation and Program Planning, 44, 1–13.
Doran, George T. (1981). There’s a S.M.A.R.T. way to write management’s goals and
objectives. Management review 70.11 1981: 35. Business Source Premier.
Frey, B. S., & Osterloh, M. (Eds.). (2001). Successful management by motivation:
Balancing intrinsic and extrinsic incentives. Springer Science & Business Media.
Gettman, H. J. (2008). Executive coaching as a developmental experience: A framework
and measure of coaching dimensions. ProQuest.
Grudens-Schuck, N. (2003). The rigidity and comfort of habits: A cultural and
philosophical analysis of the ups and downs of mainstreaming evaluation. New
Directions for Evaluation, 2003(99), 23–32.
Gudda, P. (2011). A guide to project monitoring & evaluation. Bloomington, IN:
AuthorHouse.
Harris, E., & Harvard Family Research Project (2011). Afterschool evaluation 101: How
to evaluate an expanded learning program. President and Fellows of Harvard
College Retrieved 15 December 2016 from http://www.unitedway.org/our-
impact/focus/education/out-of-school-time/tools.
Heart Safe Communities (2016). What is HEARTSafe? n.d., Retrieved 15 December
2016 from http://heartsafe-community.org/.
Hessel, V., Cortese, B., & De Croon, M. H. J. M. (2011). Novel process windows—
Concept, proposition and evaluation methodology, and intensified superheated
processing. Chemical Engineering Science, 66(7), 1426–1448.
Hofman, W. A., & Hofman, R. H. (2011). Smart management in effective schools:
Effective management configurations in general and vocational education in the
Netherlands. Educational Administration Quarterly, 47(4), 620–645. http://dx.
doi.org/10.1177/0013161X11400186.
Hummelbrunner, R. (2010). Beyond logframe: Critique, variations and alternatives.
In N. Fujita (Ed.), Beyond logframe; Using systems concepts in evaluationTokyo:
Foundation for Advanced Studies on International Development.
Isell, M. L. (2014). Health program planning and evaluation. A practical, systematic
approach for community health, 3rd ed. Jones and Bartlett.
Jung, L. A. (2007). Writing SMART objectives and strategies that fit the ROUTINE.
Teaching Exceptional Children, 39(4), 54–58.
W.K. Kellogg Foundation (2004). Logic Model Development Guide. Retrieved 15
December 2016 from https://www.wkkf.org/resource-directory/resource/
2006/02/wk-kellogg-foundation-logic-model-development-guide.
Knowlton, L. W., & Philips, C. C. (2013). The logic model guidebook: Better strategies for
great results, 2nd ed. Sage.
Lawlor, K. B. (2012). Smart goals: How the application of smart goals can contribute
to achievement of student learning outcomes. Developments in Business
Simulation and Experiential Learning, 39, .
Linstrom, J. (2006). Project objectives help you work SMART-er. Fire Chief, 50(4), 26–
28.
MacLeod, L. (2012). Making SMART goals smarter. Physician Executive, 38(2), 68–72.
Mathison, S. (2005). Encyclopedia of evaluation. Sage Publications.
Merton, R. K., & Sztompka, P. (1996). On social structure and science. University of
Chicago Press.
Montana Cardiac Ready Communities (2015). Cardiac ready community application
and toolkit. Retrieved 15 December 2016 from http://dphhs.mt.gov/
publichealth/EMSTS/cardiacready/communityapp.
North Dakota Department of Health (2016). Cardiac ready community designation
guidelines. Retrieved 15 December 2016 from https://www.health.nd.gov/
media/1382/nd-crc-criteria .
Patton, M. (2011). Developmental evaluation: Applying complexity concepts to enhance
innovation and use. New York: Guilford Press.
Pearson, E. S. (2012). Goal setting as a health behavior change strategy in overweight
and obese adults: A systematic literature review examining intervention
components. Patient Education and Counseling, 87(1), 32–42.
Picciotto, R. (2002). The logic of mainstreaming a development evaluation
perspective. Evaluation, 8(3), 322–339.
Piskurich, G. M. (2015). Rapid instructional design: Learning ID fast and right. John
Wiley & Sons.
Preskill, H., & Boyle, S. (2008). A multidisciplinary model of evaluation capacity
building. American Journal of Evaluation, 29(4), 443–459.
Renger, R., & Titcomb, A. (2002). A three-step approach to teaching logic models.
American Journal of Evaluation, 23(4), 493–503.
Renger, R. (2006). Consequences to federal programs when the logic-modeling
process is not followed with fidelity. American Journal of Evaluation, 27(4), 452–
463 2006.
Sanders, J. R. (2002). Presidential address: On mainstreaming evaluation. American
Journal of Evaluation, 23(3), 253–259.
Sharma, M., & Petosa, R. L. (2012). Measurement and evaluation for health educators.
Jones & Bartlett Publishers.
Smith, M. J. (2010). Handbook of program evaluation for social work and health
professionals. Oxford University Press.
Stevenson, J. F., Florin, P., Mills, D. S., & Andrade, M. (2002). Building evaluation
capacity in human service organizations: A case study. Evaluation and Program
Planning, 25(3), 233–243.
van der Grift, E. A., van der Ree, R., Fahrig, L., Findlay, S., Houlahan, J., Jaeger, J. A., . . .
Olson, L. (2013). Evaluating the effectiveness of road mitigation measures.
Biodiversity and Conservation, 22(2), 425–448.
Wade, D. T. (2009). Goal setting in rehabilitation: An overview of what, why and
how. Clinical Rehabilitation, 23(4), 291–295.
Williams, D. D., & Hawkes, M. L. (2003). Issues and practices related to
mainstreaming evaluation: Where do we flow from here? New Directions for
Evaluation, 2003(99), 63–83.
https://eccguidelines.heart.org/wp-content/uploads/2015/10/2015-AHA-Guidelines-Highlights-English
https://eccguidelines.heart.org/wp-content/uploads/2015/10/2015-AHA-Guidelines-Highlights-English
https://eccguidelines.heart.org/wp-content/uploads/2015/10/2015-AHA-Guidelines-Highlights-English
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0010
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0010
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0010
https://www2.ed.gov/programs/slcp/slc-wgandobj-book-f
https://www2.ed.gov/programs/slcp/slc-wgandobj-book-f
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0020
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0020
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0020
http://www.cdc.gov/dhdsp/programs/spha/evaluation_guides/docs/smart_objectives
http://www.cdc.gov/dhdsp/programs/spha/evaluation_guides/docs/smart_objectives
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0030
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0030
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0035
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0035
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0040
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0040
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0040
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0045
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0045
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0050
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0050
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0055
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0055
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0060
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0060
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0060
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0065
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0065
http://www.unitedway.org/our-impact/focus/education/out-of-school-time/tools
http://www.unitedway.org/our-impact/focus/education/out-of-school-time/tools
http://heartsafe-community.org/
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0080
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0080
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0080
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0085
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0085
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0085
http://dx.doi.org/10.1177/0013161X11400186
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0090
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0090
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0090
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0095
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0095
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0100
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0100
https://www.wkkf.org/resource-directory/resource/2006/02/wk-kellogg-foundation-logic-model-development-guide
https://www.wkkf.org/resource-directory/resource/2006/02/wk-kellogg-foundation-logic-model-development-guide
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0110
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0110
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0115
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0115
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0115
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0120
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0120
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0125
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0130
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0135
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0135
http://dphhs.mt.gov/publichealth/EMSTS/cardiacready/communityapp
http://dphhs.mt.gov/publichealth/EMSTS/cardiacready/communityapp
https://www.health.nd.gov/media/1382/nd-crc-criteria
https://www.health.nd.gov/media/1382/nd-crc-criteria
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0150
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0150
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0155
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0155
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0155
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0160
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0160
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0165
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0165
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0170
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0170
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0175
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0175
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0180
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0180
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0180
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0185
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0185
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0190
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0190
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0195
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0195
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0200
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0200
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0200
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0205
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0205
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0205
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0210
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0210
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0215
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0215
http://refhub.elsevier.com/S0149-7189(16)30258-0/sbref0215
1 Introduction
2 Case example
3 Conclusion
3.1 Lessons learned
Acknowledgements
References
SMART Goals Workgroup
7.13.23 update
Status of Previous Goals
1. Invite 4 new members to the EFC representing the
following: 1) employers, 2) providers, 3) advocates,
4) postsecondary ed, 5) Regional Center staff
a. TO INCREASE EQUITY, CREATING SELF-NOMINATION SURVEY
b. SURVEY DRAFTED FOR APPROVAL
2. Create EFC member role description
Status of Previous Goals
1. Invite 4 new members to the EFC representing the following: 1) employers, 2) providers, 3)
advocates, 4) postsecondary ed, 5) Regional Center staff
2. Create EFC member role description
a. GENERATED IN EFC INVITATION LETTER FOR APPROVAL
b. MADE CHANGES TO SHORTEN LENGTH AND USE PLAIN
LANGUAGE
New SMART Goals
•Directed by the committee in May meeting to focus
on data
•HEADLINE GOAL:
• That the Employment First Committee be a source of data
related to employment that can inform the public and
legislature regarding current status, successes, and needs.
•SMART Goal 3: Update Data Dashboard
• SPECIFIC: Update Council Data Dashboard to reflect most recently
released data, with links to expanded information
• MEASURABLE: Yes – Have the data been updated?
• ATTAINABLE: Data has been provided by DDS, identify committee
members to support, link to SCDD website supports
• RELEVANT: updated data will most accurately reflect current state of
employment
• TIME-BOUND: by October 12th meeting
•SMART Goal 4: Identify New Data Directions
• SPECIFIC: Identify 2-3 new targets for data collection that would be
informative to EFC work
• E.g., data on wait time for employment services in Regional Centers, data to
understand factors that contribute to / predict later CIE
• MEASURABLE: Yes – 2 or 3 specific ideas
• ATTAINABLE: Examine what we know and do not know, what would be
helpful to create movement, identify committee members to support
• RELEVANT: updated data will most accurately reflect employment in CA
• TIME-BOUND: by October 12th meeting propose ideas; committee can
weigh in on priorities and discuss plans for collection
How to Move Forward?
•Options:
• Formation of workgroup to pursue these two specific
SMART goals
• Formation of ongoing data workgroup
• Entire committee supports through set aside “working”
time in quarterly meetings to discuss and address these
data goals
Place an order in 3 easy steps. Takes less than 5 mins.