Energy Savings Assistance (ESA) Energy Education: Overview of Proposed Plan March 7, 2013
Dec 18, 2015
Energy Savings Assistance (ESA)Energy Education:
Overview of Proposed Plan
March 7, 2013
Agenda
2
Time Topic Lead
10:00 am Welcome / Introductions / Purpose Carol Edwards
10:15 am Overview of Research Plan Steve Westberg
10:40 am Task 1 Comprehensive ReviewValerieRichardson
11:00 am Task 2 Contractor Interviews Steve Westberg
11:20 am Task 3 Customer Qualitative In-Home Visits Steve Westberg
11:40 am Task 4 Customer Quantitative Survey Steve Westberg
12:00 pm Task 5 Savings MeasurementValerie Richardson
12:30 pm Lunch Break
1:15 pm Summary Wrap-Up / Next stepsSteve WestbergValerie Richardson
2:00 pm Adjourn All
HINER / KEMA Team: Key Members
3
Team Member Responsibilities Company
Steve Westberg Project Manager, HINER Lead (Tasks 2-4) HINER
Valerie Richardson Assistant Project Manager, KEMA Lead (Tasks 1, 5) KEMA
Rachel Schiff Comprehensive Review (Task 1) KEMA
Paul Caracciolo Field Interviews (Task 3) HINER
Luke ThelenSurvey Programming, Quantitative Analysis, and Statistics (Tasks 2-4)
HINER
Fred Coito Savings Measurement KEMA
Project Objectives
Key issues to be addressed:
• Identify best practices and potential Improvements related to HOW Energy Education is delivered (e.g., format, time, etc.)
• Identify best practices and potential Improvements related to WHAT materials and content are provided (e.g., relevance, value, gaps)
• Assess current and potential energy savings resulting from Energy Education
4
5
Project Components: Data Sources
Five data components for the study:
Stage Component Purpose
1 Comprehensive Review
(1) Guidelines, training, and materials provided to contractors (and customers). (2) IOU programs and technologies, and Out-Of-Area materials or practices that could be leveraged/adopted.
2 Contractor Interviews
(1) Understand current practices, knowledge, etc. of educators. (2) Solicit ideas for improvement from those closest to the education.
3 Customer Qualitative
Understand and explore: (1) range of educational experiences, (2) retention of content, (3) adoption of energy efficient behaviors, and (4) unmet needs.
4 Customer Quantitative
(1) Measure the prevalence of experiences, knowledge, behaviors and needs across the population of ESA participants.
5 Savings Measurement
(1) Design the experiment. (2) Employ the method, (est. completion Dec 2014)
6
Meeting Objectives
Component HOW Delivery WHAT Content Energy SavingsComprehensive Review X X (X)
Contractor Interviews X X
Customer Qualitative X X
Customer Quantitative X X (X)
Savings Measurement X
7
Meeting Objectives
Objective 1: HOW energy education is delivered
Component Research Questions
Comprehensive Review
What training has been provided to contractors? How do IOU’s assess or monitor performance? What other methods of education delivery could be employed?
Contractor Interviews
How is education delivered? What differences in delivery exist? How do customers respond? What can interfere with effective delivery?
Customer Qualitative
How has education been provided? What methods stand out? What prompts you to put learning into practice? What new methods of delivery have potential? What is missing or lacking?
Customer Quantitative How many or what percent of customers … based on qualitative issues
8
Meeting Objectives
Objective 2: WHAT materials are provided
Component Research QuestionsComprehensive Review
What materials are provided to contractors/customers? What is the specific content? What new content could be added?
Contractor Interviews
What resonates with customers? What do customers ignore? What do customers ask about that is not included?
Customer Qualitative
What content is most useful? What content does not seem to apply? What issues within the home prevent adoption? What potential new content appeals? What is missing?
Customer Quantitative How many or what percent of customers … based on qualitative issues
9
Meeting Objectives
Objective 3: Assess energy savings
Component Research QuestionsComprehensive Review
What challenges will be encountered regarding an experimental design? What sample frame and sample units are available?
Contractor Interviews --
Customer Qualitative --
Customer Quantitative To what extent have energy savings practices been implemented?
Savings Measurement
How can energy savings be measured? With what validity and reliability? What energy savings can be attributed to energy education?
10
Stage 1: Comprehensive Review
Purpose: • Document what contractors currently provide regarding Energy Education• Provide a resource for the project team to identify potential new content,
delivery methods, or additional resources
Approach Overview: • Review will include:
• Program documentation for training and delivery practices • Educational materials provided customers• Contractor implementation and supervisory practices• Third party studies and education materials • Existing and planned utility programs and technologies with potential
application to En Ed (e.g., UAT, SmartMeter tools, etc.)• Interviews with IOU managers
11
Stage 1: Comprehensive Review
Deliverables: • Written summary report that will include:
1. Listing of all materials received and reviewed2. Short description of each document or item3. Assessment by the reviewer(s) about the quality or estimated
efficacy of each document or item4. Listing of aspects of materials, training, etc. that appear to be gaps
(e.g., materials and processes that appear to be missing or lacking in the existing energy education program)
12
Stage 2: Contractor Interviews
Purpose: • Document, along with the Comprehensive Review:
• What contractors currently provide regarding energy education • The range of differences between contractors• Barriers to effective education • Best practices• Ideas for improvements from the contractors’ perspective
Approach Overview: • Contractor interviews among:
• front line (customer facing) supervisors/managers• in-home technicians
• Qualitative: In-depth telephone interviews
• Quantitative: Web-based survey
13
Stage 2: Contractor Interviews
Sampling:
Supervisor /
ManagerAssessment Technician
Installation Technician
PG&E 1 3 1
SCEOnly - 2 -
SCG Only - 2 -
SCE/SCG Joint 1 1 1
SDG&E 1 2 -
Total 3 10 2
14
Stage 2: Contractor Interviews
Discussion Topics – Managers: • How much time is allocated during an in-home visit for energy education • How do you monitor or ensure compliance in the field• What do you evaluate technicians on• How do you compensate technicians• Do you ever receive feedback from customers about the education they
have received• Etc.
15
Stage 2: Contractor Interviews
Discussion Topics – Technicians: • What training or education do you provide in homes• What are the most and least effective topics and materials• What barriers or problems that interfere with the training in the home do
you encounter (e.g., language, householder availability and interest, topics relevant to the householder)
• What type of training did you receive prior to making field visits• What was missing or lacking in your training or materials• Etc.
16
Stage 2: Contractor Interviews
Deliverables: • Written report that describes, for front-line (customer-facing)
supervisors/managers and technicians: 1. Awareness and knowledge about educational requirements and
content 2. Aspects of delivery, including time spent in total and on specific
content areas, method (e.g., walking around/demonstration, sitting at table, etc.), and recipients (e.g., homeowner, other household members)
3. Perceived obstacles or barriers to effective education4. Ideas for educational materials or delivery improvements
17
Stage 3: Customer Qualitative In-Homes
Purpose: • Compare contractor-provided information to what customers said about the
education they received to identify retention gaps• Determine what customers have retained and put into action• Identify gaps between what customers need and what they received• Identify additional opportunities for new topics, delivery methods, and
resources
18
Stage 3: Customer Qualitative
Approach Overview: • Collect feedback and input from customers, regarding:
• How existing training practices meet the needs of different households• Needs that might not be addressed• Customer motivations• How delivery and content can be improved• What information has been retained and put into practice• How household Energy Education experiences differed across
contractors and technician• What customers think about potential new education materials, content,
or resources
• 30 in-home interviews
• 6 focus groups
19
Stage 3: Customer Qualitative
Sampling:
In-Home
Interviews Focus Groups
PG&E 12 (4 clusters) 2 (1 location)
SCE/SCG 12 (4 clusters) 2 (1 location)
SDG&E 6 (2 clusters) 2 (1 location)
20
Stage 3: Customer Qualitative
Customer In-Home Interviews: • Allows us to understand the environment as well as customer experiences
and preferences. For example:• Should training be conducted during the walk-through, at the kitchen
table, or in front of a computer? • Is there value in demonstrating what the customer should do to reduce
energy use? • Is it possible to bring together all household members?
Focus Groups: • Group discussions to:
• Brainstorm ways to improve content and delivery • Review and provide feedback on new content, materials, and delivery
ideas
21
Stage 3: Customer Qualitative
Discussion Topics – In-Home Interviews: • Expectations of the program and benefits gained from reducing energy
use, knowing about gas and electric safety, etc.• Experience of receiving the educational information during the contractor’s
visits to your home: how was training conducted, etc.• What you got out of the educational information or training: what did you
learn, did you know any of this previously, have you used any of this, etc.• Barriers to implementing what you learned: habits hard to change, others
in the home not assisting, not home to take action, etc.• What might be missing from the education: do you have any unanswered
questions, any changes you would suggest for the visit
22
Stage 3: Customer Qualitative
Discussion Topics – Focus Groups: • Similar to in-home discussions, but in addition: • How can other utility programs and tools could be adopted: which of these
other programs might fit your situation, etc. • Review and brainstorm around new ideas for content and delivery of
energy education: what do you think about this idea, etc.
23
Stage 4: Customer Quantitative
Purpose:• “Validate” and quantify the qualitative findings with a telephone survey
among a representative sample of ESA participants
Approach Overview:
• 500 telephone interviews among ESA participants: • Margin of error = 4.4% at 95% confidence• Small enough to allow for recent participants to be included in the
survey
24
Stage 4: Customer Quantitative
Survey Topics: • Screening for person most involved/present during assessment visit• Household characteristics: type of home, size, age, people in the home, etc.• Measurement of attitudes, motivations, and barriers to reducing energy use• Energy education: how much time spent, information provided, etc. • Frequency of energy efficient behaviors & beliefs about success• Satisfaction with energy education• Interest in new content, programs, tools, etc. to enhance energy education • Demographics
25
Stage 4: Customer Quantitative
Deliverables:
1. Written report with percentage estimates of En Ed participants who:2. Have specific needs and motivations3. Recalled components of their energy education4. Favorable (or not) about the content and delivery of the energy education5. Implemented or took action as a result of their energy education6. Interested in and favorable toward new content and delivery ideas
26
Stage 5: Savings Measurement
Project Purpose: • Decision 12-08-044 specifies a field component to measure actual energy
savings attributable to energy education, including an experimental group of high usage (200-400% of baseline) CARE participants
Approach Overview: • Develop and implement an experimental design• Controlling the “treatment” that is delivered to customers and comparing
energy usage to a control group to isolate savings • Prior to August 31, 2013:
• Design and extract treatment and control samples from the CARE population
• Determine which energy education program elements to test • Establish monitoring and tracking process of post-treatment data after
customers receive treatment• In 2014, we will analyze a minimum of 12 months of post-treatment billing
history using either T-test and/or conducting modeling
27
Stage 5: Savings Measurement
Deliverables: • August 31, 2013:
• The framework for the experimental design including sample design and post-treatment tracking
• December 2014: • Energy savings analysis results comparing 12 months of post-
treatment behavior to control group during the same period
Updated Timeline
28
Date Task or Action
January 24 Kick-Off Meeting
February 6 Detailed Research Plan
March 7 Public Workshop
February/March Comprehensive Review
March Contractor Qualitative; Experimental Design (sample design, extraction)
April Contractor Online Quantitative; Experimental Design (test parameters)
April Customers In-Home Qualitative
May Customer Telephone Quantitative
Mid-July Data Analysis and Draft Report
Mid-August Public Workshop
August 31, 2013 Final Report
29
Summary of Comments
Review of comments provided during this meeting
30
Next Steps and Action Items
IOU’s are preparing information in response to HINER/KEMA data request
Comprehensive Review is in progress
Develop the Final Research Plan