Targeting Customers for Demand Response Based on Big Data

Targeting Customers for Demand Response Based on Big Data


Selecting customers for demand response programs is challenging and existing methodologies are hard to scale and poor in performance. The existing methods were limited by lack of temporal consumption information at the individual customer level. We propose a scalable methodology for demand response targeting utilizing novel data available from smart meters. The approach relies on formulating the problem as a stochastic integer program involving predicted customer responses. A novel approximation is developed algorithm so it can scale to problems involving millions of customers. The methodology is tested experimentally using real utility data.


smart meter data, targeting, demand response, big data, algorithms, stochastic knapsack problem.


1 Introduction

Demand response (DR) programs have been expanded to match peak load growth and increases in supply-side uncertainty. The goal of a DR program is to elicit flexibility from loads by reducing or shifting consumption in response to external signals such as prices or curtailment indicators. Typically a program is designed to extract a targeted level of energy or power from all the participating loads. The program operation yield is the ratio between the actually extracted energy from the participants and the target level. Current program yields are low, in the range of 10% to 30%.

Customer recruitment is essential for the success of demand response programs. Existing customer recruitment mechanisms are designed to identify customers that are likely to enroll. They utilize predictive marketing models (e.g. discrete choice) that utilize as inputs household related variables such as family income, enrollment in other programs and household size. The models are very successful in identifying customers that are likely to enroll, and typically 80% of recruited customers enroll. Yet, the lack of individual level consumption information has prevented these recruitment efforts from achieving high operation yields. Besides yield, a DR program needs to ensure reliable performance. Customer enrollment directly influences reliability as each consumer might provide different levels of certainty on the demand curtailment they can offer during a DR event.

The widespread deployment of advanced metering infrastructure (AMI) data can significantly change the approach to customer recruitment. This paper investigates how to target customers for DR programs based on this data. The significant recruitment cost and scalable progressive deployment of DR programs require an efficient way to select a small number of appropriate customers from a large population. Since enrollment decisions are made in advance of the actual consumption period, only a prediction of per customer DR potential is available. The prediction can be made by analyzing historical high resolution consumption data for each customer. Given a prediction, customers need to be chosen in a way that balances the magnitude of demand response potential (reward) with the uncertainty in the prediction (risk). In fact, customers need to be considered as a portfolio, and an optimal trade-off between risk and reward is desirable.

This paper develops a methodology for large-scale targeting that combines data analytics and a scalable selection procedure. The first contribution is the formulation of the customer selection problem as a stochastic discrete optimization program. This program selects the best customers in order to maximize reliability (risk) given a targeted DR amount. The second contribution is a novel guaranteed approximation algorithm for this class of discrete optimization problems which might be of independent interest. The methodology is illustrated in two climate zones of Pacific Gas & Electric, with more than 50,000 customers.

The literature on demand response programs is vast (e.g. [1], [2], [3]) and includes the definitions of DR programs [4], how to estimate the energy savings (baselining) [5] and other related areas. Recent papers have focused on designs for operating DR including distributed pricing mechanisms that adapt to data ([6], [7], [8], [9]), simplified operation mechanisms with full information ([10], [5]) and operations with partial information [11]. Integration of DR as a system reliability has also been well investigated in [12]. Finally, implementation in hardware of these solutions is being developed in [13]. Currently, most demand response targeting relies on segmentation of customers based on their monthly billing data or surveys [14], [15]. The growing availability of smart meter data has shown that such approaches are highly inaccurate since segments obtained from actual consumption differ from those obtained by alternative methods [16], [17].

The paper is organized as follows. Section 2 provides the problem description including a stochastic knapsack problem (SKP) setting and response modeling. Section 3 presents a review of SKP and develops a novel heuristic algorithm to solve the proposed SKP. Section 4 presents the experimental evaluation and validation of the methodology. Section 5 summarizes the main conclusions and discusses future work.

2 Methodology

Given the large amount of data, the scalability of the approach is very important. Thus, we propose a quick linear response modeling and a novel heuristic to solve the SKP (1), which is basically an NP-hard problem. Fig. 1 shows the overall DR program targeting flow in this paper with brief computation complexity information.

Figure 1: Overall DR program targeting flow

The remainder of the section details the methodology. It first proposes a reliability constrained selection algorithm that uses a probabilistic estimate of DR response. It then demonstrates how such response can be learnt from existing smart meter data. Notice that the specific approach to predict individual responses for different types of DR would differ, but each still provides a probabilistic estimate. For completeness, we illustrate the procedure for thermal DR.

2.1 Maximizing Demand Response Reliability

There are potential customers that can provide DR service. Given the customer recorded data, the response of customer is a random variable corresponding to the energy saved during a DR event. The distribution of is determined by fitting a response model corresponding to the type of DR and has a known joint probability distribution. The cost for customer to participate in the program is . During planning, this cost represents the cost of marketing the program to a customer and rebates for a customer to purchase the resources to perform DR. The program operator has a budget and desires an aggregate target response (in kWh) from the program with the maximum reliability possible. DR availability is captured by the control variable . DR reliability is naturally captured by the probability of the response exceeding the target . The optimal DR program selection problem can then be stated as


where represents whether a customer is recruited or not. Note that, if is same for all customers, the budget constraint is the same as limiting the number of participating customers by . The program maximizes the reliability of the DR program by recruiting customers within the program budget. The optimal reliability for budget and target is given by objective function value . The function captures the tradeoff between DR availability and DR reliability for a budget . The function has some important properties that conform to our intuition about the tradeoff. The objective function is monotonic decreasing in so if . The budget determines the constraints so if .

The proposed optimization problem is a stochastic knapsack problem (SKP). SKPs are stochastic integer programs known to be NP-hard. The goal of this paper is to develop a novel efficient approximation algorithm that scales to in millions of customers. The efficient algorithm is used then to compute the function .

An important additional assumption is that is large and is sufficiently large, so a significant number of customers are included. In that case, given a set of random variables for the response , the total group response is approximately Gaussian distributed due to the central limit theorem:


where is the recruitment vector, indicates a normal distribution, is a vector of individual response means and a covariance matrix with covariances between responses. In practice, if the number of customers selected is in the order of 50, the response distribution is very close to normal.

2.2 Response modeling

Estimating the response from provided data is an important challenge in practical applications. We illustrate here a model for estimating customer response in a specific program, but we note that our methodology applies in general, with the ability to define general models. A highlight of the methodology is that is a random variable, so models that are not accurate can still be useful.

The customer response model specification depends on the design of the demand response program. Consider a Global Temperature Adjustment (GTA) program for HVAC (air conditioning) systems [1]. Such a program increases the temperature set point of the air conditioner for each customer by a fixed amount to reduce cooling power consumption. Selecting customers with high energy saving potential during a DR event day and hour requires an accurate model to estimate the total energy consumed at each set point level. If HVAC consumption is independently metered, a simple model can be built utilizing the observed consumption, external temperature and the utilized set point [18]. In general, though, only total home consumption and external temperature are observed.

The main proportion of temperature-sensitive energy consumption is from HVAC (heating, ventilation, and air conditioning) systems [1]. It has been observed that the relationship between energy consumption and outside temperature is piecewise linear, with two or more breakpoints [5], [19], [20]. These descriptive models are typically fit utilizing past consumption data, in particular for the summer season. The power consumption of a customer at time on day , , is modeled as


where is the outside temperature, is the breakpoint temperature which is the proxy of the temperature set point for the HVAC system, is the cooling sensitivity, is the heating sensitivity or the temperature sensitivity before turning on AC system, is base load and is a random variability. Typically, in the summer is close to zero. In some cases, since either the variability is large or only large temperatures are observed in the summer, thus a reliable estimate of cannot be obtained. The model above is closely related to the ETP model [18], assuming that the system is in thermal equilibrium so inner mass temperature and outside temperature are equal.

To restrict the computational cost to find the best breakpoint, is assumed to be an integer, in the range 68-86, which is typical. Additionally, to prevent the cases that one or two dominant data points determine the breakpoint and provide invalid temperature sensitivity, we put a constraint on the breakpoint that both sides of the breakpoint should have at least certain fraction of data (e.g., we set 15% in this paper).

Model learning is performed in two steps. Minimization of residual sum of squares (RSS) is utilized to learn the parameters of the model and the distribution of the error from the observed data and . An F-test is utilized to test if to prevent overfitting. If , (3) becomes (4),


The overall computation needed to fit the consumption model is to solve (at most) 20 linear regression models: one for each potential value of the breakpoint (at most 19) and one for fitting (4). The coefficients associated with the breakpoint with smallest RSS is selected as the estimate. The regression provides estimates of parameter values and errors, in particular an estimate of cooling sensitivity and the standard deviation of parameter error denoted by . The distribution of the parameter estimate might not necessarily be Gaussian. Covariances between sensitivity estimates for different customers can be obtained by analyzing the residuals.

The DR response model is postulated as follows. The savings in the response are obtained by increasing the set point temperature by . We assume DR is utilized during very hot days, so . The response model is then postulated to be kWh. The random variable has mean and standard deviation . The covariance between any two responses and is given by . The aggregate DR response is then distributed as in (2).

3 Algorithm

In this section we design a novel scalable algorithm for customer selection that enables construction of the DR availability-reliability tradeoff.

3.1 Optimization problem transformation

Utilizing the normality assumption of the total response in the original SKP formulation (1), the following equivalence is easy to demonstrate:


The set represents the constraint set of the original SKP. The equivalency reveals some qualitative insights about the optimal solution. Consider the case , so the budget is the number of customers.If is small, then its likely the target level exceeds the group response so . The mechanism mostly selects customers with high mean, high variance responses resulting in low reliability. If is sufficiently large, then , and the mechanism selects high mean, low variance customer responses. Consequently DR reliability is high. A simple rule of thumb is that should be usually set so that the sum of the highest response means exceeds .

3.2 Previous approaches to solve the SKP

Several approaches address solving the transformed SKP in (5) ([21], [22]). The basic principle is to utilize Lemma 1 proved independently for completeness.

Lemma 1.

The customer selection SKP (1) is equivalent to the optimization program


where the constraint set is defined as

under the assumption that . Furthermore, there is a map from the optimum to . Additionally, there exists so the optimization


has the same optimum as optimization (6).


Define . Assuming there exists a so that and , the function is the same as the equivalent optimization in (5). The gradient shows monotonic decreasing in and increasing in . on the set .

By in [22], is pseudoconcave on the set. Based on the fact that a quasiconcave function over a convex set attains its minimum in an extreme point of the constraint set [23], the extreme points of should be found and tested to see which extreme point is the optimal solution. The extreme points in a convex set can be found by linear programming [24]. As shown in the gradient , when (, so and are appropriately selected), the objective function in (6) is minimized with larger and smaller . Thus, the extreme points which are feasible to be the optimal solution can be obtained by solving the problem below.

To find the extreme points in , (7) is solved multiple times for various . When is given, the calculation can be done by dynamic programming (DP). However, the overall procedure will be determined mainly by selecting a small number of various values of to find all extreme points. Setting various is related to nonlinear optimization techniques referenced in [22]. According to [25], the expected complexity of finding the set of various is (where is the number of customers, is the limit of the number of participating customers), which requires heavy computation cost when is large.

Alternatively, [26] does not try to find extreme points in , but rather changes the SKP into multiple knapsack problems (KP) on the assumption that all the s are integers and all users are independent ( is diagonal). The method is quite computationally intensive and requires solving integer LP.

A more straightforward approach is to extend the concept of a greedy approximation algorithm utilized to solve specific instances of the knapsack problem [27, 28]. The proposed algorithm is described in Appendix 6 as Algorithm 2 and has complexity . It utilizes the per customer risk-reward ratio to sort and rank customers who are offering sufficient benefit. This constraint improves the performance of the algorithm as explained in the Appendix.

3.3 Stochastic Knapsack problem solving

We propose an efficient heuristic algorithm (Algorithm 1) to solve the SKP by finding the extreme points in . By the monotonicity property shown in the proof of Lemma 1, the maximum should be obtained from one of the extreme points found by solving (7) with various . Every extreme point of has a corresponding point which can be obtained by solving


If we think a two-dimension scatter plot of (x-axis: , y-axis: ), corresponds to the slope. Then, finding an extreme point by solving (8) corresponds to finding the point on the slope with minimum intercept in . Depending on the slope, , different extreme points will be obtained. Thus, the heuristic we propose here is to find all extreme points by solving (8) constant times, , i.e., increasing the slope from to by .

0:   and from response modeling.
  Set a integer constant, (= the number of iterations).
  for  from 0 to  do
      (= equally increasing slope angle).
       Solve the problem below and save :
  end for
Algorithm 1 Algorithm to solve the SKP in (1)

We bound the ratio between the optimal cost obtained by Algorithm 1 and the true optimal cost as defined in (5):

Proposition 1.

Let . When in (5) is less than zero, Algorithm 1 has the approximation bound below, which is only depending on .


We provide the proof for Proposition 1 in Appendix 7. Briefly, the bound depends on the given data and the number of iterations, , which decide . For example, we provide the relation between and the approximation bound at Zone 13 in section 4.3 by Fig. 8.

When in (5) is larger than zero, under a given and , the optimal solution does not have to be one of the extreme points [22]. However, we assume that the optimal solution stays at one of the extreme points and find the extreme points by solving (10) instead of (9) in Algorithm 1. This is because the optimization direction is increasing and as much as possible when is larger than zero.

When is not diagonal, (9) should be solved by quadratic programming (QP) after relaxing . However, if we assume that the responses are independent so that is diagonal (, solving the problem (9) becomes a linear programming (LP) problem as shown in (12).


Especially, when the cost for each customer, , is the same, the problem simplifies to selecting the highest (at most) entries from the vector, which is the same as sorting the entries, , in a computational perspective.

To summarize, with the assumption of having independent responses and the same targeting costs, our customer selection procedure guarantees a very close to optimal solution, with a computational complexity () equivalent to that of sorting entries in a vector. This is the most significant benefit of our heuristic algorithm, which enables customer selection even with a very large number of customers.

4 Experiments on data

4.1 Description of data

The anonymized smart meter data used in this paper is provided by Pacific Gas and Electric Company (PG&E). The data contains the electricity consumption of residential PG&E customers at 1 hour intervals. There are 218,090 smart meters, and the total number of 24 hour load profiles is 66,434,179. The data corresponds to 520 different zip codes, and covers climate zones 1, 2, 3, 4, 11, 12, 13 and 16 according to the California Energy Commission (CEC) climate zone definition. The targeting methodology is tested by focusing on a cool climate zone (Zone 3: 32,339 customers) and a hot climate zone (Zone 13: 25,954 customers) during the summer season May to July 2011. Zone 3 is a coastal area and Zone 13 an inland area. Weather data corresponding to the same period is obtained from Weather Underground for each zip code in these two zones.

4.2 Consumption model fitting result

In this section, we provide analysis on the consumption model fitting result. First, the model selection result (F-test) is shown in order to check which model (between (3) and (4)) is selected for the customers depending on their climate zones. Moreover, to check whether the model explains the consumption effectively, we provide value distribution for each climate zone. As noted before, the models are aimed to capture the energy consumption of AC systems during the summer season. 25,954 household data in Zone 13 were used to fit the model, and, for comparison purposes, another 32,339 household data in Zone 3 were used.

Figure 2: Hourly model selection result by F-test
(the percentage of that model (3) is selected)

Note that the degree of freedom of model (3) is overrated because we constrained on the reference temperature to be an integer value between 68 and 86. Therefore, the F-test result will show a conservative view of the performance of the model (3). Fig. 2 shows the hourly model selection result for both climate zones. It shows the percentage of times the F-test selects model (3) for each hour. As shown in the figure, model (3) is selected much more in Zone 13 (hot) than Zone 3 (cool), which means model (3) does not perform significantly better than (4) in the cool climate zone. Also, we note that model (3) fits better in higher consumption hours (3-9PM) for both climate zones.

Considering that the system peak hours of consumption for a typical utility occur between 4PM and 6PM during the summer season, we assume that these are the hours when DR is most required. Thus, we provide the result on those hours. Fig. 3 shows the distribution during 5PM to 6PM. The other hours show almost similar results with that of 5PM to 6PM. In the hot area, the model explains about 42% of variances on average, while it only explains 10% of variances in the cool area. Considering that the only independent variable is temperature, temperature should be a very important factor in the summer season in the hot climate area, while it is not important factor of energy consumption in the cool climate area. The result is expected since cool areas tend to utilize air conditioning less frequently or even avoid it all together.

Figure 3: distribution at both climate zones

Fig. 4 shows the smoothed scatter plots of (mean response) and (response standard deviation) in (2) in both climate zones when is . Most of customers in Zone 13 have positive mean responses (coming from positive temperature sensitivity) with relatively small standard deviations, while most of the customers in Zone 3 have close to zero response, and only small fraction of customers have positive mean responses with relatively widespread standard deviations. This suggests that the explained DR program would not be as effective in the cool area as in the hot area. This is why we can achieve the same energy saving with a smaller in the hot area, as is shown in the following section 4.3.

Figure 4: and scatter plot in both climate zones

4.3 Targeting result analysis

Section 2 developed a SKP for optimal customer selection, and two different algorithms are provided to solve this combinatorial problem in Section 3.3: an efficient heuristic (Algorithm 1) and a greedy algorithm (Algorithm 2). In this section, we show the customer selection results by solving the SKP (1) with the assumptions that is same for all and is diagonal.

First, we provide the tradeoff curve between DR availability and DR reliability with fixed , which is the most important statistics for a utility manager with a limited budget to run a DR program. Then, we present the relation between varying (availability) and the minimum that achieves 95% probability. Through this plot, we can provide how much cost is required to achieve certain energy saving. Also, we show the relation between varying and the maximum probability (reliability) in (1) given , which is corresponding to the tradeoff between DR reliability and the cost with a given . Last, we provide the plot of and the approximation bound.

Figure 5: Tradeoff curve between DR availability and reliability

Fig. 5 shows the DR availability-reliability tradeoff curve within a certain limit of the number of participating customers. The interpretation example can be that, if a utility wants to save 1625kWh during 5-6PM at Zone 13, we can not guarantee the saving more than 50% with even with the best 2000 customers. A utility manager can generate this plot with setting depending on their budget and decide how much energy they will target to save with a certain reliability sacrifice.

Figure 6: Maximum probability (reliability) and N for a DR program with =

Fig. 6 shows the relation between maximum probability and given in (1). We assume that DR events can happen for each hour during 4-6PM, is and is 1000kWh. According to the plots, the heuristic algorithm always achieves equal or higher probability than the greedy algorithm. As designed, the greedy algorithm guarantees more than 50% probability when the optimal solution can achieve more than 50%. One thing to note is that the reliability changes from 0% to 100% within 30 customers in the hot climate zone though about 1100 customers are required to achieve 1000kWh of energy saving. This implies that an inappropriate number of enrollments can cause a program to fail, reinforcing the need for the analytic targeting mechanism. In contrast, in the cool climate zone, it takes more than 60 customers to reach a reliability from 0% to 100%. As shown in Fig. 4, the reason is that a small number of the customers with positive mean responses in Zone 3 have relatively large standard deviations compared to the customers in Zone 13.

Figure 7: and the minimum to achieve with more than 95% probability

Fig. 7 shows the relation between varying (availability) and the minimum required to achieve the corresponding with guaranteeing the reliability (95%). We varied T from 200kWh to 4000kWh for both climate zones. Note that energy savings of 4000kWh or more are achievable with high probability in Zone 13 while it is not possible to achieve even more than 1600kWh in Zone 3. This fact supports the reasonable conjecture again that targeting customers in hot climate zones is more effective in achieving energy savings than targeting customers in cool climate zones. Also, note that the relation between and in the figures can be approximated by quadratic equations.

Figure 8: M and approximation bound

Finally, Fig. 8 shows the approximation bound of Algorithm 1. For , we use 2000kWh, and for , the minimum that achieves more than 95% probability is selected for Zone 13. In Zone 13, from =10, the approximation bound is 0.983, which means only 11 sortings of values are needed to achieve the almost optimal solution of (1).

5 Conclusions

In this paper, we investigated an efficient customer selection process for DR programs based on a large amount of hourly consumption data. We ensured that the proposed methodology scales to large data sets and is composed of a simple linear response modeling and a fast heuristic selection algorithm to solve the SKP. Moreover, we proved the approximation bound of the heuristic algorithm solution, which is an almost optimal solution in the hot climate zone.

As an example DR program showing the result of our approach, we suggested a DR program controlling the temperature set point during the summer season. From the experimental results, we found that there are many customers, even in cool climate zones, who are very sensitive to the outside temperature and use their AC system actively. Therefore, when the number of recruiting customers, , is small, the energy saving potential can be higher in the cool climate zone. The probability of achieving the targeted energy savings changes rapidly over a small range of in hot climate zones, which means it is very important to select the right customers to achieve the targeted energy savings with minimum enrollment and cost for the utility side.

The proposed method can be extended in many different ways. For example, the proposed heuristic algorithm does not have to be confined to a DR program targeting. If any application problem can be formulated in the same optimization problem format, the heuristic algorithm can be applied with low computational cost. Also, by changing DR programs and response modeling, more refined and practical research can be done. Additionally, it might be important to include other practical constraints or requirements from DR programs.


6 Gradual greedy approximation

We propose a very simple greedy algorithm to solve (1). Excluding the design parameter from (5), we minimize with the assumption that is diagonal. We sort customers by descending order of . However, when is larger than the sum of for the first customers, naive sorting can be far from the optimal solution. For example, when the s are very small even with a high ratio, it cannot make smaller than zero even though is smaller than zero. To prevent this drawback, we can modify the greedy approximation algorithm to the gradual greedy algorithm, Algorithm 2. This greedy algorithm will select one best customer at a time while satisfying the condition in (3). Therefore, the algorithm guarantees that it achieves at least more than 50 when in (5). When , we let the greedy algorithm select the customers with the biggest .

0:   and vectors in (12).
  ; .
  for  from 1 to  do
     Find from below and set =1 in :
  end for
Algorithm 2 Gradual Greedy Algorithm

7 Proof for Proposition 1

Figure 9: When optimal point is not captured in Algoritham 1

Let is the minimum obtained by Algorithm 1. And, we denote the optimal value of by in (5). Briefly, if the optimal point is one of the extreme points found by our heuristic algorithm (Algorithm 1), the algorithm achieves the actual optimal solution. However, as seen in Fig. 9, the optimal point can be in the triangular region defined by the two consecutive slopes () and the two corresponding extreme points. Then, the optimal point satisfies the properties below.


Then, we can derive the inequality as seen in (7).


From the proof, the approximation bound is given by the minimum ratio between two consecutive . It would be closer to 1 as is larger, which means Algorithm 1 can achieve a solution within a very tight bound from the optimal solution. The relation between and the approximation bound on the actual data is provided in Fig. 8 in Section 4.3.


This research was funded in part by the Department of energy ARPA-E under award number DE-AR0000018, the California Energy Commission under award number PIR-10-054, and Precourt Energy Efficiency Center. The views and opinions of the authors expressed herein do not necessarily state or reflect those of the United States Government or any agency thereof.

We acknowledge the support of PGE for providing the anonymized data and in particular Brian Smith for his comments on this work. We also want to thank Sam Borgeson and Adrian Albert for their efforts to create the weather database.


  1. J. Han and M. A. Piette, “Solutions for summer electric power shortages: Demand response and its applications in air conditioning and refrigerating systems,” Lawrence Berkeley National Laboratory, 2008.
  2. S. Borenstein, M. Jaske, and A. Rosenfeld, “Dynamic pricing, advanced metering, and demand response in electricity markets,” Center for the Study of Energy Markets, UC Berkeley, 2002.
  3. M. Albadi and E. El-Saadany, “Demand response in electricity markets: An overview,” in IEEE Power Engineering Society General Meeting, vol. 2007, pp. 1–5, 2007.
  4. F. Rahimi and A. Ipakchi, “Demand response as a market resource under the smart grid paradigm,” Smart Grid, IEEE Transactions on, vol. 1, no. 1, pp. 82–88, 2010.
  5. J. L. Mathieu, P. N. Price, S. Kiliccote, and M. A. Piette, “Quantifying changes in building electricity use, with application to demand response,” Smart Grid, IEEE Transactions on, vol. 2, no. 3, pp. 507–518, 2011.
  6. N. Li, L. Chen, and S. H. Low, “Optimal demand response based on utility maximization in power networks,” in Power and Energy Society General Meeting, 2011 IEEE, pp. 1–8, IEEE, 2011.
  7. Q. Dong, L. Yu, W.-Z. Song, L. Tong, and S. Tang, “Distributed demand and response algorithm for optimizing social-welfare in smart grid,” in Parallel & Distributed Processing Symposium (IPDPS), 2012 IEEE 26th International, pp. 1228–1239, IEEE, 2012.
  8. C. Chen, S. Kishore, Z. Wang, M. Alizadeh, and A. Scaglione, “How will demand response aggregators affect electricity markets?—a cournot game analysis,” in Communications Control and Signal Processing (ISCCSP), 2012 5th International Symposium on, pp. 1–6, IEEE, 2012.
  9. A. D. Dominguez-Garcia and C. N. Hadjicostis, “Distributed algorithms for control of demand response and distributed energy resources,” in Decision and Control and European Control Conference (CDC-ECC), 2011 50th IEEE Conference on, pp. 27–32, IEEE, 2011.
  10. H. Zhong, L. Xie, and Q. Xia, “Coupon incentive-based demand response: Theory and case study,” Power Systems, IEEE Transactions on, vol. 28, no. 2, pp. 1266–1276, 2013.
  11. J. A. Taylor, J. L. Mathieu, D. S. Callaway, and K. Poolla, “Price and capacity competition in zero-mean storage and demand response markets,” in Communication, Control, and Computing (Allerton), 2012 50th Annual Allerton Conference on, pp. 1316–1323, IEEE, 2012.
  12. L. Xie, P. M. Carvalho, L. A. Ferreira, J. Liu, B. H. Krogh, N. Popli, and M. D. Ilic, “Wind integration in power systems: Operational challenges and possible solutions,” Proceedings of the IEEE, vol. 99, no. 1, pp. 214–232, 2011.
  13. E.-K. Lee, R. Gadh, and M. Gerla, “Energy service interface: Accessing to customer energy resources for smart grid interoperation,” Selected Areas in Communications, IEEE Journal on, vol. 31, no. 7, pp. 1195–1204, 2013.
  14. L. Lutzenhiser et al., “Behavioral assumptions underlying california residential energy efficiency programs,” CIEE Energy & Behavior Program, 2009.
  15. S. J. Moss, M. Cubed, and K. Fleisher, “Market segmentation and energy efficiency program design,” Berkeley, California Institute for Energy and Environment, 2008.
  16. J. Kwac, J. Flora, and R. Rajagopal, “Household energy consumption segmentation using hourly data,” Smart Grid, IEEE Transactions on, vol. 5, no. 1, pp. 420–430, 2014.
  17. B. A. Smith, J. Wong, and R. Rajagopal, “A simple way to use interval data to segment residential customers for energy efficiency and demand response program targeting,” ACEEE Proceedings, 2012.
  18. R. C. Sonderegger, “Dynamic models of house heating based on equivalent thermal parameters,” Ph.D. dissertation, Princeton Univ., 1978.
  19. B. J. Birt, G. R. Newsham, I. Beausoleil-Morrison, M. M. Armstrong, N. Saldanha, and I. H. Rowlands, “Disaggregating categories of electrical energy end-use from whole-house hourly data,” Energy and Buildings, vol. 50, pp. 93–102, 2012.
  20. J. R. Cancelo, A. Espasa, and R. Grafe, “Forecasting the electricity load from one day to one week ahead for the spanish system operator,” International Journal of Forecasting, vol. 24, no. 4, pp. 588–602, 2008.
  21. A. M. Geoffrion, “Solving bicriterion mathematical programs,” Operations Research, vol. 15, no. 1, pp. 39–54, 1967.
  22. M. I. Henig, “Risk criteria in a stochastic knapsack problem,” Operations Research, vol. 38, no. 5, pp. 820–825, 1990.
  23. M. S. Bazaraa, H. D. Sherali, and C. M. Shetty, Nonlinear programming: theory and algorithms. John Wiley & Sons, 2013.
  24. S. P. Boyd and L. Vandenberghe, Convex optimization. Cambridge university press, 2004.
  25. M. I. Henig, “The shortest path problem with two objective functions,” European Journal of Operational Research, vol. 25, no. 2, pp. 281–291, 1986.
  26. D. P. Morton and R. K. Wood, On a stochastic knapsack problem and generalizations, vol. 9. Springer, 1998.
  27. G. B. Dantzig, “Discrete-variable extremum problems,” Operations Research, vol. 5, no. 2, pp. 266–288, 1957.
  28. B. C. Dean, M. X. Goemans, and J. Vondrdk, “Approximating the stochastic knapsack problem: The benefit of adaptivity,” in Foundations of Computer Science, 2004. Proceedings. 45th Annual IEEE Symposium on, pp. 208–217, IEEE, 2004.
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters
Comments 0
The feedback must be of minumum 40 characters
Add comment

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question