As electric vehicles (EVs) roll into garages, highways, and loading docks across the United States, the power grid is facing a significant transformation. While most drivers are focused on how far their next charge will take them, grid operators are asking an equally important question: when and where is all this new EV charging demand going to hit the electric system?
Anticipating and managing these new loads is crucial. If done well, it could help utilities avoid costly and potentially unnecessary grid expansions. If left unaddressed, utilities risk being caught off guard by sudden spikes in electric demand, leading to overloaded feeders, service disruptions, and expensive last-minute infrastructure upgrades.
That is why the Pennsylvania Public Utility Commission (PA PUC) tasked DSA with developing a 15-year granular forecast of the impact of EV adoption as part of a statewide study on transmission and distribution (T&D) capacity. These forecasts can allow utilities to plan for future changes in demand while also identifying opportunities to defer or avoid infrastructure upgrades.
| Figure 1: Forecasted HE19 EV Demand across Pennsylvania in 2030 | |
| Figure A: PECO
|
Figure B: PPL
|
| Figure C: Duquesne
|
Figure D: FirstEnergy
|
We developed a granular, location-specific EV load forecast across the four major electric utility territories in Pennsylvania:
Step 1 – Forecast Statewide EV Energy Consumption
We used PJM’s zonal EV energy consumption forecasts to develop a 15-year projection of statewide EV energy consumption specific to the service territories of each of Pennsylvania’s four major electric distribution companies (EDCs).
Step 2 – Forecast EV Energy Consumption by Charging Type
We divided the total energy forecast into three main types of EV charging: light-duty vehicle (LDV) at home charging, LDV public/workplace charging, and medium- and heavy-duty vehicle (MHDV) charging. Each charging type was assigned a different hourly load shape; the LDV shapes were borrowed from NREL and the MHDV shape was borrowed from PJM’s electric vehicle forecast.
Step 3 – Estimate Feeder Propensity
After mapping feeders to locations, each feeder on the grid was assigned a propensity score for EV adoption and the three types of EV charging:
Step 4 – Calibrate to the Statewide Forecast
Using a calibration function developed by DSA, the forecasted EV growth for each utility was distributed across that utility’s feeders, ensuring that local adoption is calibrated to every location’s propensity and adds up to the utility-wide and statewide total forecasts.
Step 5 – Estimate Hourly Loads
Annual feeder-level EV MWh were converted to hourly loads using the normalized load shapes for each charging type. This enables planners to see not just how much demand will grow, but at what time during the day loads will peak on each circuit.
| Figure 2: Example of Projected Hourly EV Load for a Duquesne Feeder in 2025 and 2030 |
![]() |
This type of granular forecasting is key to identify T&D upgrade deferral opportunities or target T&D upgrades. Instead of making system-wide upgrades, utilities can use forecasts like these to invest where upgrades are really needed and avoid investments where they are not.
As Pennsylvania and other states prepare for rapid growth in electric vehicles, location-specific T&D forecasting will be essential to making smart, cost-effective infrastructure investment decisions.
Demand Side Analytics has done multiple HER evaluations for many utilities; across a range of geographies, fuels, and cohort sizes. In this post, we review the results of a persistence study conducted in Pennsylvania after the conclusion of report delivery at four of FirstEnergy’s Pennsylvania Electric Distribution Companies (EDCs)[1]. The effect of treatment is typically measured through comparison of the group of customers receiving the HER, known as the treatment group, to a statistically identical control group. The comparison is done both in the period prior to receiving the reports (to assess treatment and control equivalence), and after treatment (to measure the impact of treatment on consumption). The goal of the study was to identify how long energy savings persisted, even after reports were discontinued.
What it is
Behavioral conservation programs, such as residential Home Energy Reports (HERs), are well understood to provide small, yet measurable reductions in energy use when appropriately deployed. These programs are relatively inexpensive, geographically widespread, and effective at reducing consumption for most residential customer segments. The treatment effect is related to behavior changes brought about by providing customers information about their energy consumption relative to their peers. By showing how much energy the customer is using compared to similar households, the HER induces behavior changes using the power of social norms. This effect is facilitated by having the HER provide energy efficiency and conservation tips to the customer, which induces temporary and permanent behavior changes. Because of this, the conservation effect can persist in treated groups even after customers stop receiving reports.
Why is it important
The persistence of HER impacts means that even after the discontinuation of HER delivery, treated customers continue to provide energy savings relative to customers who never received a report. Accurately quantifying how long savings persist in the previously-treated group is important as it helps determine program cost-effectiveness and assessments of the effective useful life of any HER program.
How did we do the analysis
The HER program in question was implemented as a randomized control trial for each of the EDCs. A randomized control trial is an evaluation technique that provides very precise and unbiased estimates of the effect of treatment – that is, the receipt of HER bill comparisons. If properly implemented, randomized control trials (RCTs) are a very effective framework for estimating HER impacts for two key reasons, related to how HER programs are designed:
To isolate the impact of treatment while controlling other factors that may influence energy use, DSA applied a lagged dependent regression approach. This model works particularly well at providing precise savings estimates when there is good pretreatment equivalence between the treatment and control groups. The model uses information about individual household seasonal consumption patterns collected through billing data analysis to estimate the impact of treatment in each month after the start of report delivery, including after reports were stopped for the persistence test.
To model the effect of persistence, a simple regression specification was used to determine the decay of impacts as a function of the number of months since the cohort received their last report. Because impacts can be seasonal and have uncertainty around them, a weighted average of the prior year’s monthly impacts was used to create an average pre-cessation savings level.
The key metric used to quantify the effect of persistence is how many months it takes for impacts to reach zero. Once the regression is performed, DSA used the intercept and slope from the regression output to calculate the number of months it would take for the trend in impacts to go to zero. This is shown graphically below, where it takes approximately 37 months for the orange trend line to cross the y-axis at zero. The intercept for the persistence regression line is set equal to the average savings in the prior 12-months (shown in blue circles and the grey squares at month = 0). The underlying assumption with this model is that the HER savings will continue to decay at the same rate observed in months 1-24 until reaching zero.
Figure 1: Persistence Modeling Example

Program-specific considerations
Each EDC studied in this project had multiple cohorts of customers that were included in the HER program and persistence study. Not all of these cohorts showed robust pretreatment equivalence. Because of this, it is best to carefully consider which cohort’s impacts should be included in an analysis of HER persistence. The criteria that DSA used to categorize cohort quality were threefold:
These criteria are illustrated graphically in Figure 2 for one of the EDCs. The x-axis plots the average number of customers still active in the period between June 2016 and May 2018 for each cohort, while the y-axis shows the percentage of the original cohort size that is still active during this period. The markers for each cohort are also color-coded to highlight whether the cohort was used in the final analysis, or what the reason was for its exclusion.
Figure 2: Cohort Characterization for Met-Ed

What are the results
The cohort characterization resulted in five cohorts analyzed in the persistence study: two from Met-Ed and three from Penelec. The five cohorts that qualified were then fed into a second-stage model that sought to determine the monthly decay rate of the savings estimates. Since there is noise in each savings estimate and seasonal variation in the savings estimates, DSA thought it most appropriate to set the intercept of each cohort’s regression to equal the average savings percentage over the twelve months immediately prior to the persistence test. That is, the starting point of this regression was not simply what the customers saved in May of 2016 but a weighted average of the full year prior to the test. Figure 3 shows the raw data used to construct this analysis. The five cohorts that were identified as having good equivalence and the appropriate cohort size are shown in the figure below. The trend line of persistent savings is shown in blue. This figure displays the trend for FirstEnergy cohorts only and approaches zero nearly 30 months after the HER reports stop being sent to customers. This estimate is combined with other Pennsylvania studies, below, to provide an overall decay rate estimate.
Figure 3: FirstEnergy Persistence Trends

To estimate the HER effect duration more precisely, DSA fit a simple linear model that related the percent savings estimates – again weighted by the aggregate reference load – to the number of months it had been since the cohort received a HER. The weighting of the percent savings is necessary in this case because we are using percent savings as our variable of interest. Doing the weighting ensures that larger cohorts are have more impact than smaller ones, and that a 2% savings in a high-consumption month counts more than a 2% savings in a low-consumption month, while still creating a percentage metric that can be directly compared to other studies.
Table 8:Persistence Trends by Cohort

How do these results compare to a larger set of recent HER persistence studies?
In 2015, the Pennsylvania evaluation team conducted a similar analysis of residential HER persistence for cohorts from PPL and Duquesne Energy that stopped receiving HERs. Three cohorts across these two EDCs experienced between 16 and 24 months of no report delivery, with resumption of HERs after that period had passed. Prior to having begun the persistence test, the two PPL cohorts had received reports since 2010 (Legacy), and since 2011 (Expansion). Duquesne’s HER program began in PY4 (between June 2012 and May 2013), so at most customers received 11 months of HER treatment prior to report discontinuation.
Table 9:Persistence Trends for Other Pennsylvania HER Studies

In general, the FirstEnergy results are quite similar to those of the two PPL cohorts, with between 29.7 to 51 months of expected impact decay time. The PPL customers in the HER program had been receiving reports for a longer period than most FirstEnergy customers, but had generally similar savings rates prior to the start of the persistence test. This generally corresponds to the common understanding of HER reports; namely that they can deliver relatively consistent savings after a maturation period of one to two years when customers first start receiving reports. The decay rates, or slope of percent savings decay, in the PPL study is quite similar to that of FirstEnergy, with between a 0.04% and 0.06% drop in savings per month (roughly a 0.5% to 0.75% annual decay).
[1] The full report can be found here: http://www.puc.pa.gov/Electric/pdf/Act129/SWE_Res_Behavioral_Program-Persistence_Study_Addendum2018.pdf
Demand Side Analytics (DSA) conducted a transmission and distribution (T&D) avoided cost study in Pennsylvania. The focus of the study is on quantifying the change in T&D costs associated with an increase or decrease of kW coincident with location-specific peaks. It employs methodologies that are novel for Pennsylvania but have been applied and approved in New York and California for load forecasting and distributed energy resource (DER) valuation. The full study is available online here.
A vital role of the Electric Distribution Company (EDC) is to ensure that regional electricity supply makes its way to homes and businesses safely, reliably, and cost-effectively. By projecting future demand and reinforcing the local transmission and distribution network so that sufficient capacity is available to meet local needs as they change over time, costly outages are avoided.
The study was designed to meet the following objectives:
The deferral value approach focuses on quantifying the value of load relief on ratepayer costs (i.e. revenue requirements). It effectively compares revenue requirements with and without load relief. While infrastructure upgrades can be temporarily avoided or deferred via load relief, they cannot be avoided indefinitely because equipment eventually ages and needs to be replaced. The marginal cost of service study approach quantifies the supply cost of additional distribution or transmission capacity on the system. At the simplest level, it involves classifying infrastructure investments as growth related or not and dividing the costs of those investments by the incremental transmission or distribution capacity added. The approach uses the cost of adding additional transmission capacity to the system as a proxy for the cost avoided by reducing peak demand.
Figure 1: T&D Avoided Costs Methods Considered

Figure 2 describes the main steps in developing location-specific avoided distribution costs using probabilistic methods. These steps help identify the magnitude of reductions at the right location at the right time and right season to delay upgrades. The process was implemented for each feeder and substation (transformers or terminals if applicable) that had a valid growth rate and operating limit, then layered two levels to get the distribution avoided cost for each site. For system-wide values, the estimates consider the likelihood that reductions would be in locations with or without value due to random chance. We emphasize that system-wide value is a load-weighted average of areas where reductions do or do not lead to deferral of distribution investments.
Figure 2: Key Steps in Estimating Location Specific Avoided Costs

Based on the analysis, Table 1 summarizes the territory-wide average T&D avoided cost by EDC and season. The avoided cost of transmission capacity estimates simply increases with inflation and escalation. The avoided cost of distribution capacity values changes at varying rates based on the outcomes of the probabilistic deferral analysis methodology.
Table 1: T&D Avoided Cost ($2026)

While the final study outputs are territory-wide average values for each EDC, the granular forecasts are useful for identifying locations and timing when demand reductions or injections of distributed generation are beneficial. Figure 3 shows the total deferral value by local systems. Shades of blue indicate relatively low deferral value while orange and red tones indicate high deferral value. The values range from a lower bound of $100 to a maximum of $200 in 2026 nominal dollar. To calculate the total deferral value, we aggregated the deferral value at feeder and substation levels, then incorporated the deferral value of transmission. A key outcome of the study was to highlight the fact that the avoided T&D costs associated with peak load relief vary widely within each EDC territory.
The value of avoided T&D costs associated with an increment or decrement of peak load is a key component of benefit-cost analyses. In practice, T&D capital costs resources are concentrated in pockets that are experiencing growth but lack the capacity to accommodate additional growth. Most utilities have a mix of areas where loads are growing and areas where loads are declining, which may or may not overlap with highly loaded components. In locations with excess distribution capacity or where local peak demand is declining, the potential to avoid T&D costs is minimal. In areas where a large, growth‐related investment is imminent, the avoided T&D costs from reducing peak demand are much higher.
Figure 3: Heat Map of Total Deferral Value ($2026/kW-year)

Load growth varies by location. Some pockets are experiencing load growth, and some are experiencing load decreases. We received granular growth rates for PECO, PPL, and FirstEnergy. In each EDC territory, growth trends varied by location. As a result, growth-related T&D investments are required even when overall EDC loads are flat or declining.
The T&D avoided costs are concentrated in locations that are more heavily loaded. A key component of distribution planning is the loading factor: the weather-normalized peak demand divided by the operating limit. Not surprisingly, avoided costs are concentrated in more highly loaded locations. Conversely, locations with ample capacity to accommodate additional loads had lower avoided T&D costs.
Individual locations are generally winter or summer peaking, not both. Most distribution locations – feeders, transformers, substations – can be classified as winter or summer peaking. Few feeders are dual peaking. The implication is that the avoidable T&D cost for a specific location is concentrated in the summer or winter, but not both.
Resources that deliver load relief at the right location, in the right season, and at the right hours are more valuable. The same energy efficiency resource can deliver different T&D benefits at two locations based on how well it coincides with the local peak load. To illustrate, a more efficient air conditioner does not provide T&D load relief on a winter-peaking substation but does so on a summer-peaking substation. Likewise, measures with load shapes that better coincide with the need for load relief are more valuable.
Lump loads are a key driver of distribution upgrades. Lump loads are simply new, large loads. They vary widely in size, and it can be difficult to predict in advance when and where they will show up. When they are built, they often trigger distribution and even transmission upgrades. Reducing demand via EE&C program efforts can create room for additional loads and help avoid upgrades due to smaller lump-load projects.
The Pennsylvania PUC leveraged the results of the study for its 2026 TRC Test Tentative Order. The TRC Test Order provides utilities with directions for calculating avoided costs and performing benefit-cost analysis when planning for Phase V of Act 129 programs. The avoided T&D values will play an important role in the Phase V DR Potential Study, which DSA is currently working on.
Demand Side Analytics (DSA) recently conducted two similar studies on the accuracy of using smart meter data for evaluation and settlement of energy efficiency (meter-based methods). Different localities refer to these meter-based methods with distinctive terminology. Across our two studies, California[1] (for Pacific Gas & Electric) refers to these methods as normalized metered energy consumption (NMEC) while Vermont[2] (for the Vermont Department of Public Service) refers to them as Advanced Measurement & Verification (M&V).
While these studies were distinct, they were both concerned, to varying degrees, with:
What follows is a review of the benefits of using meter-based methods for estimating EE program impacts, a brief overview of each study’s goals, and our findings.
The primary challenge of estimating energy savings is the need to accurately detect changes in energy consumption due to the energy efficiency intervention, while systematically eliminating plausible alternative explanations for those changes. Did the introduction of energy efficiency measures cause a change in energy use? Or can the differences be explained by other factors (such as the effects of the COVID-19 pandemic)? To evaluate energy savings, it is necessary to estimate what energy consumption would have been in the absence of program intervention—the counterfactual or baseline.
Meter-based methods rely on whole-building, site-specific electric and/or gas consumption data, either at the hourly or daily level. This data is used to estimate energy savings associated with the installation of individual or multiple energy efficiency measures (EEMs) at the site.
Many methods exist to estimate savings associated with EEMs, all with varying degrees of modeling complexity, data requirements, accuracy, and precision. The benefits of using meter-based methods include:
Pacific Gas and Electric Company (PG&E) currently uses the CalTRACK Version 2.0 method (CalTRACK) to estimate avoided energy use for its energy efficiency programs based on the Population-Level NMEC methodology. A notable feature of the population NMEC method has been the lack of comparison groups, which are used to adjust the energy savings baseline and normalize the savings estimate for factors beyond weather. The pre-post method without a comparison group relies almost exclusively on weather normalization and effectively assumes that the only difference between the pre- and post-intervention periods is weather and the installation of EEMs. The COVID-19 pandemic laid bare the limitations of the adopted method. The pandemic led to changes in our commutes, business operations, and home use patterns. Not surprisingly, it has also changed how, when, and how much electricity and gas we use. Moreover, the impact on energy use differs for residential customers and various types of businesses.
Given the changes in energy consumption that have occurred over the course of the COVID-19 pandemic, the need for alternative approaches to CalTRACK and similar, simple pre-post regression methods for estimating EE impacts is paramount. While adding comparison groups typically improves the accuracy of these energy saving estimates, there are three main logistical challenges:
To determine if there are viable alternative models that can accommodate the effects of the COVID-19 pandemic or other wide-scale non-routine events, DSA conducted an accuracy assessment of the existing Population NMEC methods as well as a variety of other methods with and without comparison groups.
Accurate and unbiased estimates of energy efficiency impacts are critical for utility program staff, third-party program implementers, and regulators. In evaluating the accuracy of the existing Population NMEC methods used in the PG&E territory, we tested a variety of other methods, with and without comparison groups, to simulate a competition and identify the methods that are unbiased and accurate (Figure 1).
The accuracy of these methods are assessed by applying placebo treatment on customers that did not participate in EE programs during the period analyzed. The impact of a program (or in this case, a pseudo-program) is calculated by estimating a counterfactual and comparing it to the observed consumption during the post-treatment period. Because no EEMs were installed in this simulation, any deviation between the counterfactual and actual loads is due to error. The process is repeated hundreds of times – a procedure known as bootstrapping – to construct the distribution of errors.
Figure 1: General Approach for Accuracy Assessment

Figure 2: Distribution of Error across Comparison Groups

Given these findings, rather than try to produce a single prescriptive method for NMEC analyses of energy efficiency programs, we instead recommend a framework by which proposed NMEC methods can be tested, certified, and used to estimate savings:
The primary objective of the Hourly Impact of Energy Efficiency Evaluation Pilot was to better understand the time-value of energy efficiency measure savings and the implications for program design, delivery, and evaluation. Because energy efficiency in the Northeast qualifies for capacity value, accurate estimates of the contribution of energy efficiency to peak hours is critical. Using high-frequency 15-minute consumption data from Green Mountain Power’s AMI and program tracking data from Efficiency Vermont, the study team modeled energy consumption of participating homes and businesses separately in the pre-installation and the post-installation periods. These two periods were compared to understand how consumption changed following installation of an energy efficiency or beneficial electrification measure. A secondary objective of the study was to compare Advanced M&V methods, or regression-based modeling of utility meter data, with the approaches traditionally used in Vermont. This comparison helped to determine where Advanced M&V could offer cost savings, improve the accuracy and granularity of savings estimates, and identify lessons for program operations.
To generate savings for the 21 prescriptive measures and the 124 custom projects in Vermont, we implement Advanced M&V procedures that build upon the International Performance Measurement and Verification Protocol (IPMVP) Option C Whole Facility approach to energy savings estimation. We do this through a regression model that follows Lawrence Berkeley National Laboratory’s (LBNL) Time-of-Week Temperature (TOWT) Model, where the dependent variable is hourly electric consumption from the meter and the independent variables contain information about the weather, day of week, and time of day.
This methodology estimates efficiency impacts in each hour of the year. Granular results provide insight into the distribution of energy savings across a year. For example, Figure 3 shows a heat map of the average energy savings from installing a variable speed heat pump. This measure’s model estimates a large load increase during the winter months (blue regions). Negative savings is a good thing in this case because it means Vermont homes are using the heat pump for heating and displacing delivered fuel consumption. There is also a pocket of denser load increase in the summer months during the middle of the day (orange regions), presumably due to homes that may not have had air conditioning previously using the heat pump as an air conditioner.
Figure 3: Variable Speed Heat Pump Heat Map

Figure 4: Example of a Well-Behaved Custom Project
Given these findings, to have a chance at accurately and precisely estimating savings from efficiency measures, the guidance below must be taken into consideration: