CPUC Workshop on Best Practices & Lessons Learned in Time Variant Pricing TVP Pilot Design and Load Impact M&V Dr. Stephen George Senior Vice President.

Slides:



Advertisements
Similar presentations
Load Impacts And Marketing Effectiveness from The Countrys Largest Dynamic Pricing Program Load Impacts And Marketing Effectiveness from The Countrys Largest.
Advertisements

SmartPOWER Critical Peak Pricing (CPP) June 3, 2008.
Energy Analysis Department Electricity Markets and Policy Group DOE Smart Grid Investment Grant Program: Dynamic Pricing & Consumer Behavior Studies Chuck.
Load Impact Estimation for Demand Response Resources Nicole Hopper, Evaluation Manager July 14, 2009 National Town Meeting on Demand Response and Smart.
Time-of-Use and Critical Peak Pricing
1 The Potential For Implementing Demand Response Programs In Illinois Rick Voytas Manager, Corporate Analysis Ameren Services May 12, 2006.
2013 Statewide BIP Load Impact Evaluation Candice Churchwell DRMEC Spring 2014 Load Impacts Evaluation Workshop San Francisco, California May 7, 2014.
2013 SDG&E Summer Saver Load Impact Evaluation Dr. Stephen George DRMEC Spring 2014 Load Impacts Evaluation Workshop San Francisco, California May 7, 2014.
SMUD’s SmartPricing Options Marketing Strategy Jennifer Potter July 31, 2014 Powering forward. Together.
Overview of ARRA-Funded SGIG Consumer Behavior Studies Peter Cappers (LBNL) A National Town Meeting on Demand Response and Smart Grid July 11, LBNL.
1 SMUD’s Small Business Summer Solutions Pilot: Behavioral response of small commercial customers to DR programs (with PCTs) Karen Herter, Ph.D. Associate.
Chapter 12 - Forecasting Forecasting is important in the business decision-making process in which a current choice or decision has future implications:
February 23, 2006Karen Herter, LBL/CEC/UCB-ERG 1 /29 Temperature Effects on Residential Electric Price Response Karen Herter February 23, 2006.
1 PG&E’s Operating Experience with TVP Rates Best Practices and Lessons Learned in Time-Variant Pricing R Residential Rate Workshop Gregory B.
Types of Studies Observational Study –Observes and measures characteristics without trying to modify the subjects being studied Experiment –Impose a treatment.
Part 3 of 3 By: Danielle Davidov, PhD & Steve Davis, MSW, MPA INTRODUCTION TO RESEARCH: SAMPLING & DESIGN.
The Efficiency of Energy Efficiency Program Tom Van Paris Vice President-Member Services & Communications October 18, 2012.
Presentation Overview
California Statewide Pricing Pilot Lessons Learned Roger Levy Demand Response Research Center NARUC Joint Meeting Committee on Energy.
Colorado Rural Electric Association Energy Innovations Summit Demand Response: Are Customers Ready to Change Their Ways? Confidential October 27, 2014.
Learning Objective Chapter 8 Primary Data Collection: Experimentation CHAPTER eight Primary Data Collection: Experimentation Copyright © 2000 by John Wiley.
Measurement, Verification, and Forecasting Protocols for Demand Response Resources: Chuck Goldman Lawrence Berkeley National Laboratory.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
SmartRate Best Practices and Lessons Learned
TRANSLATING RESEARCH INTO ACTION What is Randomized Evaluation? Why Randomize? J-PAL South Asia, April 29, 2011.
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Non-Experimental Methods Florence Kondylis.
CPUC Residential Rate Rulemaking Workshop July 31, 2014 SDG&E TVP Marketing and Outreach Proposals.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Experimental Design making causal inferences Richard Lambert, Ph.D.
Chapter Four Experimental & Quasi-experimental Designs.
EvergreenEcon.com ESA 2011 Impact Evaluation Draft Report Public Workshop #2 August 7, 2013 Presented By: Steve Grover, President.
Demand Response and the California Information Display Pilot 2005 AEIC Load Research Conference Myrtle Beach, South Carolina July 11, 2005 Mark S. Martinez,
BPA M&V Protocols Overview of BPA M&V Protocols and Relationship to RTF Guidelines for Savings and Standard Savings Estimation Protocols.
2013 California Statewide Critical Peak Pricing Evaluation Josh L. Bode Candice A. Churchwell DRMEC Spring 2014 Load Impacts Evaluation Workshop San Francisco,
Linking the Wholesale and Retail Markets through Dynamic Retail Pricing Presented by: Henry Yoshimura Manager, Demand Response ISO New England September.
CPUC Workshop on Best Practices & Lessons Learned in Time Variant Pricing TVP Load & Bill Impacts, Role of Technology & Operational Consideration Dr. Stephen.
BGE Limited Income Pilot Programs - Evaluation ACI Home Performance Conference March 2012.
Study Session Experimental Design. 1. Which of the following is true regarding the difference between an observational study and and an experiment? a)
Market research for a start-up. LEARNING OUTCOMES By the end of this lesson I will be able to: –Define and explain market research –Distinguish between.
California’s Proposed DR Cost-Effectiveness Framework January 30, 2008.
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
Natural Experiments Versus RCTs Steven Gortmaker, Ph.D. Harvard School of Public Health /Harvard Prevention Research Center.
1 Module 3 Designs. 2 Family Health Project: Exercise Review Discuss the Family Health Case and these questions. Consider how gender issues influence.
California’s Statewide Pricing Pilot Summer 2003 Impact Evaluation 17 th Annual Western Conference, San Diego, California Ahmad Faruqui and Stephen S.
DR issues in California discussed last year in March Historical DR in California: some background issues –Twenty years of programs/tariffs I/C and AC cycling.
4.4 Marketing Research.
Utilities’ Update on Energy Savings Assistance Program Studies Ordered in D LIOB Meeting August 21, 2013 Sacramento, California.
Randomized Assignment Difference-in-Differences
Evaluation Research Dr. Guerette. Introduction Evaluation Research – Evaluation Research – The purpose is to evaluate the impact of policies The purpose.
Government’s Evolving Role in Resource Planning and Environmental Protection Arthur H. Rosenfeld, Commissioner California Energy Commission April 19, 2002.
Chapter Nine Primary Data Collection: Experimentation and
Building Blocks for Premise Level Load Reduction Estimates ERCOT Loads in SCED v2 Subgroup July 21, 2014.
1 Proposed Policies to Increase the level of Demand Response Energy Action Plan Update April 24 th, 2006, Sacramento, CA Mike Messenger, CEC.
CEC Public Workshop Order Instituting Informational and Rulemaking Proceeding (08-DR-01) March 3, 2008.
2015 California Statewide Critical Peak Pricing Evaluation DRMEC Spring 2016 Load Impact Evaluation Workshop San Francisco, California May, 2016 Prepared.
You’re Getting Warmer... Smart Thermostat Pilot Update Presenters: Dan Rubado-ETO, Evaluation Project Manager Dennis Rominger-PSE, Market Manager Customer.
2013 Load Impact Evaluation Capacity Bidding Program (CBP) Steve Braithwait, Dan Hansen, and Dave Armstrong Christensen Associates Energy Consulting DRMEC.
Pay-As-You-Go Update November Program Overview Paying advance for electricity Flexibility and Customer benefits Make as many payments as necessary.
DRMEC Spring 2016 Load Impacts Evaluation Workshop San Francisco, California May 10, SDG&E Summer Saver Load Impact Evaluation.
2015 SDG&E PTR/SCTD Evaluation DRMEC Spring 2016 Load Impact Workshop George Jiang May 11 th, 2016 Customer Category Mean Active Participants Mean Reference.
2013 Load Impact Evaluation of Southern California Edison’s Peak Time Rebate Program Josh Schellenberg DRMEC Spring 2014 Load Impact Evaluation Workshop.
Pay-As-You-Go Final 2012 Report. Agenda PAYG Refresher Pilot Goals & Overview Voice of the Customer Front Office Impacts Back Office Impacts Financial.
Introducing Smart Energy Pricing Cheryl Hindes
Right-sized Evaluation
Highlights from SMUD's SmartPricing Options Pilot
Workshop Presentation
Resource Adequacy Demand Forecast Coincidence Adjustments
Impact Evaluation Methods
Behavior Modification Report with Peak Reduction Component
Presentation transcript:

CPUC Workshop on Best Practices & Lessons Learned in Time Variant Pricing TVP Pilot Design and Load Impact M&V Dr. Stephen George Senior Vice President Nexant June 17, 2014

 With RCTs, customers are randomly assigned to treatment and control conditions, so the only difference between the groups (except for random chance) is the treatment itself  Impacts are calculated as the difference between usage for the treatment and control groups after the treatment goes into effect minus the difference between the two groups before the treatment goes into effect (to eliminate any small difference due to change) 1 The gold standard of experimental designs is a randomized control trial (RCT) or randomized encouragement design (RED) Treatment Group (T) Control Group (C) kWh post pre post kWh Pretreatment Post Treatment Impact = Difference-in-difference = (Tpost – Cpost) – (Tpre – Cpre)

 With an RED, customers are randomly assigned to two groups and one group (the encouraged group) is offered the treatment and the other (the control group) is not – not everyone offered the treatment takes it  Impacts are calculated in a two step process 2 An RED has internal validity equal to that of an RCT Group Offered Treatment (T)) Control Group Not Offered Treatment (C) kWh post pre post kWh Pretreatment Post Treatment Intention-to-Treat Impact = Difference-in-difference = (Tpost – Cpost) – (Tpre – Cpre) Step 1: Intention to Treat Impact Step 2: Local Average Treatment Effect (LATE) Divide Intention-to-Treat estimate by the % of customers offered the treatment who accepted it

 For opt-in programs, RCTs require a “recruit and delay” or “recruit and deny” strategy to eliminate selection bias  Choosing two random groups of customers, offering the treatment to one and comparing usage for those who took it to those not offered it is NOT a valid experimental design (although it is often done)  Utilities are often concerned about customer dissatisfaction with recruit and delay or deny, but these methods have been used very successfully ●SMUD’s SPO employed recruit and delay (for two years) for two treatments without any significant customer backlash – and surveys show no difference in customer satisfaction with SMUD between customers who were and were not delayed ●PG&E used recruit and deny for its recent Honeywell/Opower PCT pilot ●Marblehead Municipal Light used recruit and delay for its TVP pilot  RCTs are much easier to implement for default pilots and programs (e.g., Home Energy Reports)  If opt-out rates are large, RCTs morph into REDs and should be analyzed as a RED -- all technology programs are defacto REDs because of installation issues 3 RCTs and REDs Have Advantages & Disadvantages

 For event based options (e.g., CPP tariffs, load control, etc.), RCTs can effectively fly below the radar as long as it is possible to rotate who is called for each event – PG&E uses a large RCT to estimate impacts with extremely high accuracy for the SmartAC program 4 RCTs and REDs Have Advantages & Disadvantages There are 10 lines here, all extremely similar with each based on a sample of roughly 12,000 SmartAC customers

 REDs are typically much simpler to implement than RCTs because you don’t have to delay or deny recruitment to anyone but they can also be more expensive because of the need for larger sample sizes  If enrollment rates and/or expected impacts are small, quite large sample sizes may be required to estimate the intention-to-treat effect  10% enrollment with 10% average impacts means you need to estimate a 1% effect, which requires many thousands of observations 5 RCTs and REDs Have Advantages & Disadvantages

 With enough time, money and willpower, most things of interest can be explored through a well constructed pilot  California (and the industry) need several more pilots testing default and opt-in enrollment side-by-side  SMUD and ComEd came out with very different results  If legal, it would be great to implement one or more additional default pilots in the next couple of years as input to developing long term pricing strategy – it is NOT appropriate to extrapolate average impact estimates based on opt-in pilots to default deployments  Most industry pilots have focused more on load impacts than customer acceptance  Additional research on customer preferences for different rate options would be useful  How best to market TVP options can be determined by conducting “mini pilots” within the context of a controlled launch 6 What Research Questions Can TVP Pilots Help Answer?

7 It is Possible to Conduct “Mini Pilots” or Tests of Marketing Options In the Early Stages of Program Launch It’s possible to conduct dozens of test of marketing options within a few weeks or months of project launch by simultaneously conducting multiple tests on small random samples of customers – this can support a multi-stage cycle of continuous improvement in marketing and cost effectiveness

 For existing tariffs and programs, RCTs and REDs typically are not possible  Although, as discussed before, it may be possible to use an RCT for event-based tariffs if notification can be controlled and alternated)  Quasi-experimental methods are typically used to estimate impacts  “Within-subjects” methods estimate impacts based on reference loads for individual customers developed from periods when the tariff is not in effect (e.g., prior to the customer going on the tariff or, for dynamic rates, on “event like” days when events are not called)  Another approach is to develop a comparison (control) group based on statistical matching methods (e.g., propensity score matching) – these methods match participants with non- participants that have similar usage patterns and other observable characteristics during the periods when the tariffs are not in effect and then estimate impacts in the same was as for an RCT or RED  Both methods rely on the availability of suitable pretreatment data – in the absence of suitable data, both approaches can produce biased results  With the right data, statistical matching eliminates the need to model weather and other factors that cause usage to fluctuate from one day to the next 8 What Are the Primary Methods to Measure End-Use Conservation and Peak Demand Reduction for TVP Rates?

For comments or questions, contact: Stephen George Senior Vice President, Utility Services Nexant, Inc. 101 Montgomery St., 15 th Floor San Francisco, CA