Tag Archives: M.Cubed

The State Water Board needs to act to start Flood MAR pilot projects

I recently presented to CDWR’s Lunch-MAR group the findings for a series of studies we conducted on the universe of benefits from floodwater managed aquifer recharge (MAR) and the related economic and financing issues. I also proposed that an important next step is to run a set of pilots to study the acceptability of on-farm floodwater recharge projects to growers, including how do they respond to incentives and program design, and what are the potential physical consequences.

The key to initiating these pilots is getting a clear declaration from the State Water Resources Control Board that excess floodwaters are surplus and available. Unfortunately, the Water Board has not provided sufficient clarification on how these projects can take “advantage of seasonal or occasional flood waters that overtop the banks of a stream and are then directed into a designated recharge area.” Instead, the Board’s website says that such diverted floodwaters cannot be stored for future beneficial use–which obviates the very purpose of retaining the floodwaters in the first place.

The Board should be at least issuing temporary use permits for floodwaters above certain designated levels as being available for pilot projects on the basis that non-use of those floodwaters constitute a surrender of that right for the year. Then those agencies interested in flood MAR can design projects to experiment with potential configurations.

Can Net Metering Reform Fix the Rooftop Solar Cost Shift?: A Response

A response to Severin Borenstein’s post at UC Energy Institute where he posits a large subsidy flowing to NEM customers and proposes an income-based fixed charge as the remedy. Borenstein made the same proposal at a later CPUC hearing.

The CPUC is now considering reforming the current net energy metering (NEM) tariffs in the NEM 3.0 proceeding. And the State Legislature is considering imposing a change by fiat in AB 1139.

First, to frame this discussion, economists are universally guilty of status quo bias in which we (since I’m one) too often assume that changing from the current physical and institutional arrangement is a “cost” in an implicit assumption that the current situation was somehow arrived at via a relatively benign economic process. (The debate over reparations for slavery revolve around this issue.) The same is true for those who claim that NEM customers are imposing exorbitant costs on other customers.

There are several issues to be considered in this analysis.

1) In looking at the history of the NEM rate, the emergence of a misalignment between retail rates that compensate solar customers and the true marginal costs of providing service (which are much more than the hourly wholesales price–more on that later) is a recent event. When NEM 1.0 was established residential rates were on the order of 15 c/kWh and renewable power contracts were being signed at 12 to 15 c/kWh. In addition, the transmission costs were adding 2 to 4 c/kWh. This was the case through 2015; NEM 1.0 expired in 2016. NEM 2.0 customers were put on TOU rates with evening peak loads, so their daytime output is being priced at off peak rates midday and they are paying higher on peak rates for usage. This despite the fact that the difference in “marginal costs” between peak and off wholesale costs are generally on the order of a penny per kWh. (PG&E NEM customers also pay a $10/month fixed charge that is close to the service connection cost.) Calculating the net financial flows is more complicated and deserve that complex look than what can be captured in a simple back of the envelope calculation.

2) If we’re going to dig into subsidies, the first place to start is with utility and power plant shareholders. If we use the current set of “market price benchmarks” (which are problematic as I’ll discuss), out of PG&E’s $5.2 billion annual generation costs, over $2 billion or 40% are “stranded costs” that are subsidies to shareholders for bad investments. In an efficient marketplace those shareholders would have to recover those costs through competitively set prices, as Jim Lazar of the Regulatory Assistance Project has pointed out. One might counter those long term contracts were signed on behalf of these customers who now must pay for them. Of course, overlooking whether those contracts were really properly evaluated, that’s also true for customers who have taken energy efficiency measures and Elon Musk as he moves to Texas–we aren’t discussing whether they also deserve a surcharge to cover these costs. But beyond this, on an equity basis, NEM 1.0 customers at least made investments based on an expectation, that the CPUC did not dissuade them of this belief (we have documentation of how at least one county government was mislead by PG&E on this issue in 2016). If IOUs are entitled to financial protection (and the CPUC has failed to enact the portfolio management incentive specified in AB57 in 2002) then so are those NEM customers. If on the other hand we can reopen cost recovery of those poor portfolio management decisions that have led to the incentive for retail customers to try to exit, THEN we can revisit those NEM investments. But until then, those NEM customers are no more subsidized than the shareholders.

3) What is the true “marginal cost”? First we have the problem of temporal consistency between generation vs. transmission and distribution grid (T&D) costs. Economists love looking at generation because there’s a hourly (or subhourly) “short run” price that coincides nicely with economic theory and calculus. On the other hand, those darn T&D costs are lumpy and discontinuous. The “hourly” cost for T&D is basically zero and the annual cost is not a whole lot better. The current methods debated in the General Rate Cases (GRC) relies on aggregating piecemeal investments without looking at changing costs as a whole. Probably the most appropriate metric for T&D is to calculate the incremental change in total costs by the number of new customers. Given how fast utility rates have been rising over the last decade I’m pretty sure that the “marginal cost” per customer is higher than the average cost–in fact by definition marginal costs must be higher. (And with static and falling loads, I’m not even sure how we calculated the marginal costs per kwh. We can derive the marginal cost this way FERC Form 1 data.) So how do we meld one marginal cost that might be on a 5-minute basis with one that is on a multi-year timeframe? This isn’t an easy answer and “rough justice” can cut either way on what’s the truly appropriate approximation.

4) Even if the generation cost is measured sub hourly, the current wholesale markets are poor reflections of those costs. Significant market distortions prevent fully reflecting those costs. Unit commitment costs are often subsidized through out of market payments; reliability regulation forces investment that pushes capacity costs out of the hourly market, added incremental resources–whether for added load such as electrification or to meet regulatory requirements–are largely zero-operating cost renewables of which none rely on hourly market revenues for financial solvency; in California generators face little or no bankruptcy risk which allows them to underprice their bids; on the flip side, capacity price adders such as ERCOT’s ORDC overprices the value of reliability to customers as a backdoor way to allow generators to recover investments through the hourly market. So what is the true marginal cost of generation? Pulling down CAISO prices doesn’t look like a good primary source of data.

We’re left with the question of what is the appropriate benchmark for measuring a “subsidy”? Should we also include the other subsidies that created the problem in the first place?

Is the NASDAQ water futures market transparent enough?

Futures markets are settled either physically with actual delivery of the contracted product, or via cash based on the difference in the futures contract price and the actual purchase price. The NASDAQ Veles California Water Index future market is a cash settled market. In this case, the “actual” price is constructed by a consulting firm based on a survey of water transactions. Unfortunately this method may not be full reflective of the true market prices and, as we found in the natural gas markets 20 years ago, these can be easily manipulated.

Most commodity futures markets, such at the crude oil or pork bellies, have a specific delivery point, such as Brent North Sea Crude or West Texas Intermediate at Cushing, Oklahoma or Chicago for some livestock products. There is also an agreed upon set of standards for the commodities such as quality and delivery conditions. The problem with the California Water Index is that these various attributes are opaque or even unknown.

Two decades ago I compiled the most extensive water transfer database to date in the state. I understand the difficulty of collecting this information and properly classifying it. The bottom line is that there is not a simple way to clearly identify what is the “water transfer price” at any given time.

Water supplied for agricultural and urban water uses in California has many different attributes. First is where the water is delivered and how it is conveyed. While water pumped from the Delta gets the most attention, surface water comes from many other sources in the Sacramento and San Joaquin Valleys, as well as from the Colorado River. The cost to move this water greatly varies by location ranging from gravity fed to a 4,000 foot lift over the Tehachapis.

Second is the reliability and timing of availability. California has the most complex set of water rights in the U.S. and most watersheds are oversubscribed. A water with a senior right delivered during the summer is more valuable than a junior right delivered in the winter.

Third is the quality of the water. Urban districts will compete for higher quality sources, and certain agricultural users can use higher salinity sources than others.

A fourth dimension is that water transfers are signed for different periods and delivery conditions as well as other terms that directly impact prices.

All of these factors lead to a spread in prices that are not well represented by a single price “index”. This becomes even more problematic when a single entity such as the Metropolitan Water District enters the market and purchases one type of water which they skews the “average.” Bart Thompson at Stanford has asked whether this index will reflect local variations sufficiently.

Finally, many of these transactions are private deals between public agencies who do not reveal key attributes these transfers, particularly price, because there is not an open market reporting requirement. A subsequent study of the market by the Public Policy Institute of California required explicit cooperation from these agencies and months of research. Whether a “real time” index is feasible in this setting is a key question.

The index managers have not been transparent about how the index is constructed. The delivery points are not identified, nor are the sources. Whether transfers are segmented by water right and term is not listed. Whether certain short term transfers such as the State Water Project Turnback Pool are included is not listed. Without this information, it is difficult to measure the veracity of the reported index, and equally difficult to forecast the direction of the index.

The housing market has many of these same attributes, which is one reason why you can’t buy a house from a central auction house or from a dealer. There are just too many different dimensions to be considered. There is housing futures market, but housing has one key difference from the water transfer market–the price and terms are publicly reported to a government agency (usually a county assessor). Companies such as CoreLogic collect and publish this data (that is distributed by Zillow and Redfin.)

In 2000, natural gas prices into California were summarized in a price index reported by Natural Gas Intelligence. The index was based a phone survey that did not require verification of actual terms. As part of the electricity crisis that broke that summer, gas traders found that they could manipulate gas prices for sales to electricity generators higher by simply misreporting those prices or by making multiple sequential deals that ratcheted up the price. The Federal Energy Regulatory Commission and Commodity Futures Trading Commission were forced to step in and establish standards for price reporting.

The NASDAQ Veles index has many of the same attributes as the gas market had then but perhaps with even less regulatory protections. It is not clear how a federal agency could compel public agencies, including the U.S. Bureau of Reclamation, to report and document prices. Oversight of transactions by water districts is widely dispersed and usually assigned to the local governing board.

Trying to introduce a useful mechanism to this market sounds like an attractive option, but the barriers that have impeded other market innovations may be too much.

What is driving California’s high electricity prices?

This report by Next10 and the University of California Energy Institute was prepared for the CPUC’s en banc hearing February 24. The report compares average electricity rates against other states, and against an estimate of “marginal costs”. (The latter estimate is too low but appears to rely mostly on the E3 Avoided Cost Calculator.) It shows those rates to be multiples of the marginal costs. (PG&E’s General Rate Case workpapers calculates that its rates are about double the marginal costs estimated in that proceeding.) The study attempts to list the reasons why the authors think these rates are too high, but it misses the real drivers on these rate increases. It also uses an incorrect method for calculating the market value of acquisitions and deferred investments, using the current market value instead of the value at the time that the decisions were made.

We can explore the reasons why PG&E’s rates are so high, much of which is applicable to the other two utilities as well. Starting with generation costs, PG&E’s portfolio mismanagement is not explained away with a simple assertion that the utility bought when prices were higher. In fact, PG&E failed in several ways.

First, PG&E knew about the risk of customer exit as early as 2010 as revealed during the PCIA rulemaking hearings in 2018. PG&E continued to procure as though it would be serving its entire service area instead of planning for the rise of CCAs. Further PG&E also was told as early as 2010 (in my GRC testimony) that it was consistently forecasting too high, but it didn’t bother to correct thee error. Instead, service area load is basically at the save level that it was a decade ago.

Second, PG&E could have procured in stages rather than in two large rounds of request for offers (RFOs) which it finished by 2013. By 2011 PG&E should have realized that solar costs were dropping quickly (if they had read the CEC Cost of Generation Report that I managed) and that it should have rolled out the RFOs in a manner to take advantage of that improvement. Further, they could have signed PPAs for the minimum period under state law of 10 years rather than the industry standard 30 years. PG&E was managing its portfolio in the standard practice manner which was foolish in the face of what was occurring.

Third, PG&E failed to offer part of its portfolio for sale to CCAs as they departed until 2018. Instead, PG&E could have unloaded its expensive portfolio in stages starting in 2010. The ease of the recent RPS sales illustrates that PG&E’s claims about creditworthiness and other problems had no foundation.

I calculated the what the cost of PG&E’s mismanagement has been here. While SCE and SDG&E have not faced the same degree of exit by CCAs, the same basic problems exist in their portfolios.

Another factor for PG&E is the fact that ratepayers have paid twice for Diablo Canyon. I explain here how PG&E fully recovered its initial investment costs by 1998, but as part of restructuring got to roll most of its costs back into rates. Fortunately these units retire by 2025 and rates will go down substantially as a result.

In distribution costs, both PG&E and SCE requested over $2 billion for “new growth” in each of its GRCs since 2009, despite my testimony showing that growth was not going to materialize, and did not materialize. If the growth was arising from the addition of new developments, the developers and new customers should have been paying for those additions through the line extension rules that assign that cost responsibility. The utilities’ distribution planning process is opaque. When asked for the workpapers underlying the planning process, both PG&E and SCE responded that the entirety were contained in the Word tables in each of their testimonies. The growth projections had not been reconciled with the system load forecasts until this latest GRC, so the totals of the individual planning units exceeded the projected total system growth (which was too high as well when compared to both other internal growth projections and realized growth). The result is a gross overinvestment in distribution infrastructure with substantial overcapacity in many places.

For transmission, the true incremental cost has not been fully reported which means that other cost-effective solutions, including smaller and closer renewables, have been ignored. Transmission rates have more than doubled over the last decade as a result.

The Next10 report does not appear to reflect the full value of public purpose program spending on energy efficiency, in large part because it uses a short-run estimate of marginal costs. The report similarly underestimates the value of behind-the-meter solar rooftops as well. The correct method for both is to use the market value of deferred resources–generation, transmission and distribution–when those resources were added. So for example, a solar rooftop installed in 2013 was displacing utility scale renewables that cost more than $100 per megawatt-hour. These should not be compared to the current market value of less than $60 per megawatt-hour because that investment was not made on a speculative basis–it was a contract based on embedded utility costs.

The PCIA is heading California toward another energy crisis

The California ISO Department of Market Monitoring notes in its comments to the CPUC on proposals to address resource adequacy shortages during last August’s rolling blackouts that the number of fixed price contracts are decreasing. In DMM’s opinion, this leaves California’s market exposed to the potential for greater market manipulation. The diminishing tolling agreements and longer term contracts DMM observes is the result of the structure of the power cost indifference adjustment (PCIA) or “exit fee” for departed community choice aggregation (CCA) and direct access (DA) customers. The IOUs are left shedding contracts as their loads fall.

The PCIA is pegged to short run market prices (even more so with the true up feature added in 2019.) The PCIA mechanism works as a price hedge against the short term market values for assets for CCAs and suppresses the incentives for long-term contracts. This discourages CCAs from signing long-term agreements with renewables.

The PCIA acts as an almost perfect hedge on the retail price for departed load customers because an increase in the CAISO and capacity market prices lead to a commensurate decrease in the PCIA, so the overall retail rate remains the same regardless of where the market moves. The IOUs are all so long on their resources, that market price variation has a relatively small impact on their overall rates.

This situation is almost identical to the relationship of the competition transition charge (CTC) implemented during restructuring starting in 1998. Again, energy service providers (ESPs) have little incentive to hedge their portfolios because the CTC was tied directly to the CAISO/PX prices, so the CTC moved inversely with market prices. Only when the CAISO prices exceeded the average cost of the IOUs’ portfolios did the high prices become a problem for ESPs and their customers.

As in 1998, the solution is to have a fixed, upfront exit fee paid by departing customers that is not tied to variations in future market prices. (Commissioner Jesse Knight’s proposal along this line was rejected by the other commissioners.) By doing so, load serving entities (LSEs) will be left to hedging their own portfolios on their own basis. That will lead to LSEs signing more long term agreements of various kinds.

The alternative of forcing CCAs and ESP to sign fixed price contracts under the current PCIA structure forces them to bear the risk burden of both departed and bundled customers, and the IOUs are able to pass through the risks of their long term agreements through the PCIA.

California would be well service by the DMM to point out this inherent structural problem. We should learn from our previous errors.

“What are public benefits of conveyance?” presented to the California Water Commission

Maven’s Notebook posted a summary of presentations to the California Water Commission by Richard McCann of M.Cubed, Steve Hatchett of Era Economics, and David Sunding of the Brattle Group. Many of my slides are included.

The Commission is developing a framework that might be used to identify how shares of conveyance costs might be funded by the state of California. The Commission previously awarded almost $3 billion in bond financing for a dozen projects under the Proposition 1B Water Storage Investment Program (WSIP). That process used a prescribed method including a Technical Guide that determined the eligible public benefits for financing by the state. M.Cubed supported the application by Irvine Ranch Water District and Rio Bravo-Rosedale Water Storage District for the Kern Fan water bank.

We’ve already paid for Diablo Canyon

As I wrote last week, PG&E is proposing that a share of Diablo Canyon nuclear plant output be allocated to community choice aggregators (CCAs) as part of the resolution of issues related to the Integrated Resource Plan (IRP), Resource Adequacy (RA) and Power Charge Indifference Adjustment (PCIA) rulemakings. While the allocation makes sense for CCAs, it does not solve the problem that PG&E ratepayers are paying for Diablo Canyon twice.

In reviewing the second proposed settlement on PG&E costs in 1994, we took a detailed look at PG&E’s costs and revenues at Diablo. Our analysis revealed a shocking finding.

Diablo Canyon was infamous for increasing in cost by more than ten-fold from the initial investment to coming on line. PG&E and ratepayer groups fought over whether to allow $2.3 billion dollars.  The compromise in 1988 was to essentially shift the risk of cost recovery from ratepayers to PG&E through a power purchase agreement modeled on the Interim Standard Offer Number 4 contract offered to qualifying facilities (but suspended as oversubscribed in 1985).

However, the contract terms were so favorable and rich to PG&E, that Diablo costs negatively impacted overall retail rates. These costs were a key contributing factor that caused industrial customers to push for deregulation and restructuring. As an interim solution in 1995 in anticipation of forthcoming restructuring, PG&E and ratepayer groups arrived at a new settlement that moved Diablo Canyon back into PG&E’s regulated ratebase, earning the utilities allowed return on capital. PG&E was allowed to keep 100% of profit collected between 1988 and 1995. The subsequent 1996 settlement made some adjustments but arrived at essentially the same result. (See Decision 97-05-088.)

While PG&E had borne the risks for seven years, that was during the plant startup and its earliest years of operation.  As we’ve seen with San Onofre NGS and other nuclear plants, operational reliability is most at risk late in the life of the plant. PG&E’s originally took on the risk of recovering its entire investment over the entire life of the plant.  The 1995 settlement transferred the risk for recovering costs over the remaining life of the plant back to ratepayers. In addition, PG&E was allowed to roll into rate base the disputed $2.3 billion. This shifted cost recovery back to the standard rate of depreciation over the 40 year life of the NRC license. In other words, PG&E had done an end-run on the original 1988 settlement AND got to keep the excess profits.

The fact that PG&E accelerated its investment recovery over the first seven years and then shifted recovery risk to ratepayers implies that PG&E should be allowed to recover only the amount that it would have earned at a regulated return under the original 1988 settlement. This is equal to the discounted net present value of the net income earned by Diablo Canyon, over both the periods of the 1988 (PPA) and 1995 settlements.

In 1996, we calculated what PG&E should be allowed to recover in the settlement given this premise.  We assumed that PG&E would be allowed to recover the disputed $2.3 billion because it had taken on that risk in 1988, but the net income stream should be discounted at the historic allowed rate of return over the seven year period.  Based on these assumptions, PG&E had recovered its entire $7.7 billion investment by October 1997, just prior to the opening of the restructured market in March 1998.  In other words, PG&E shareholders were already made whole by 1998 as the cost recovery for Diablo was shifted back to ratepayers.  Instead the settlement agreement has caused ratepayers to pay twice for Diablo Canyon.

PG&E has made annual capital additions to continue operation at Diablo Canyon since then and a regulated return is allowed under the regulatory compact.  Nevertheless, the correct method for analyzing the potential loss to PG&E shareholders from closing Diablo is to first subtract $5.1 billion from the plant in service, reducing the current ratebase to capital additions incurred since 1998. This would reduces the sunk costs that are to be recovered in rates from $31 to $3 per megawatt-hour.

Note that PG&E shareholders and bondholders have earned a weighted return of approximately 10% annually on this $5.1 billion since 1998. The compounded present value of that excess return was $18.1 billion by 2014 earned by PG&E.

Microgrids could cost 10% of undergrounding PG&E’s wires

One proposed solution to reducing wildfire risk is for PG&E to put its grid underground. There are a number of problems with undergrounding including increased maintenance costs, seismic and flooding risks, and problems with excessive heat (including exploding underground vaults). But ignoring those issues, the costs could be exorbitant-greater than anyone has really considered. An alternative is shifting rural service to microgrids. A high-level estimate shows that using microgrids instead could cost less than 10% of undergrounding the lines in regions at risk. The CPUC is considering a policy shift to promote this type of solution and has new rulemaking on promoting microgrids.

We can put this in context by estimating costs from PG&E’s data provided in its 2020 General Rate Case, and comparing that to its total revenue requirements. That will give us an estimate of the rate increase needed to fund this effort.

PG&E has about 107,000 miles of distribution voltage wires and 18,500 in transmission lines. PG&E listed 25,000 miles of distribution lines being in wildfire risk zones. The the risk is proportionate for transmission this is another 4,300 miles. PG&E has estimated that it would cost $3 million per mile to underground (and ignoring the higher maintenance and replacement costs). And undergrounding transmission can cost as much as $80 million per mile. Using estimates provided to the CAISO and picking the midpoint cost adder of four to ten times for undergrounding, we can estimate $25 million per mile for transmission is reasonable. Based on these estimates it would cost $75 billion to underground distribution and $108 billion for transmission, for a total cost of $183 billion. Using PG&E’s current cost of capital, that translates into annual revenue requirement of $9.1 billion.

PG&E’s overall annual revenue requirement are currently about $14 billion and PG&E has asked for increases that could add another $3 billion. Adding $9.1 billion would add two-thirds (~67%) to PG&E’s overall rates that include both distribution and generation. It would double distribution rates.

This begs two questions:

  1. Is this worth doing to protect properties in the affected urban-wildlands interface (UWI)?
  2. Is there a less expensive option that can achieve the same objective?

On the first question, if we look the assessed property value in the 15 counties most likely to be at risk (which includes substantial amounts of land outside the UWI), the total assessed value is $462 billion. In other words, we would be spending 16% of the value of the property being protected. The annual revenue required would increase property taxed by over 250%, going from 0.77% to 2.0%.

Which turns us to the second question. If we assume that the load share is proportionate to the share of lines at risk, PG&E serves about 18,500 GWh in those areas. The equivalent cost per unit for undergrounding would be $480 per MWh.

The average cost for a microgrid in California based on a 2018 CEC study is $3.5 million per megawatt. That translates to $60 per MWh for a typical load factor. In other words a microgrid could cost one-eighth of undergrounding. The total equivalent cost compared to the undergrounding scenario would be $13 billion. This translates to an 8% increase in PG&E rates.

To what extent should we pursue undergrounding lines versus shifting to microgrid alternatives in the WUI areas? Should we encourage energy independence for these customers if they are on microgrids? How should we share these costs–should locals pay or should they be spread over the entire customer base? Who should own these microgrids: PG&E or CCAs or a local government?

 

 

 

 

End the fiction of regulatory oversight of California’s generation

1542383922472

M.Cubed is the only firm willing to sign the non-disclosure agreements (NDA) that allow us to review the investor-owned utilities’ (IOUs) generation portfolio data on behalf of outside intervenors, such as the community choice aggregators (CCAs). Even the direct access (DA) customers who constitute about a quarter of California’s industrial load are represented by a firm that is unwilling to sign the NDAs. This situation places departed load customers, and in fact all customers, at a distinct disadvantage when trying to regulate the actions of the IOUs. It is simply impossible for a single small firm to scrutinize all of the filings and data from the IOUs. (Not to mention that one, SDG&E, gets a complete free pass for now as that it has no CCAs.)

This situation has arisen because the NDAs require that the “reviewing representatives” not be in a position to advise market participants, such as CCAs or energy service providers (ESPs) that sell to DA customers, on procurement decisions. This is an outgrowth of AB 57 in 2002, a state law passed to bring IOUs back into the generation market after the collapse of restructuring in 2001. That law was intended to the balance of power to the IOUs away from generators for procurement purposes. Now it puts the IOUs at a competitive advantage against other load serving entities (LSEs) such as CCAs and ESPs, and even bundled customers.

This imbalance has arisen for several insurmountable reasons:

  • No firm can build its business on serving only to review IOU filings without offering other procurement consulting services to clients.
  • It is difficult to build expertise for reviewing IOU filings without participating in procurement services for other LSEs or resource providers. (I am uniquely situated by the consulting work I did for the CEC on assessing generation technology costs for over a decade.)
  • CPUC staff similarly lacks the expertise for many of the same reasons, and are relatively ineffective at these reviews. The CPUC is further limited by its ability to recruit sufficient qualified staff for a variety of reasons.

If California wants to rein in the misbehavior by IOUs (such as what I’ve documented on past procurement and shareholder returns earlier), then we have two options to address this problem going forward:

  1. Transform at least the power generation management side of the IOUs into publicly owned entities with more transparent management review.
  2. End the annual review and setting of PCIA and CTC rates by establishing one-time prepayment amounts. By prepaying or setting a fixed annual amount, the impact of accounting maneuvers are diminished substantially, and since IOUs can no longer shift portfolio management risks to departed load customers, the IOUs more directly face the competitive pressures that should make them more efficient managers.

Moving forward on Flood-MAR with pilots

The progress on implementing floodwater managed aquifer recharge programs (Flood-MAR) reminds me of the economist’s joke, “sure it works in practice, but does it work in theory?” A lot of focus seems to be on trying to refine the technical understanding of recharge, without going with what we already know about aquifer replenishment from decades of applications.

The Department of Water Resources Flood-MAR program recently held a public forum to discuss its research program. I presented a poster (shown above) on the findings of a series of studies we conducted for Sustainable Conservation on the economic and financial considerations for establishing these programs. (I posted about this last February.)

My conclusion from the presentations and the other publications we’ve followed is that the next step is to set up pilots using different institutional set ups and economic incentives. The scientists and engineers can further refine their findings, but we generally know where the soils are better for percolation versus others, and we know that crop productivity won’t fall too much where fields are flooded. The real issues fall into five categories, of which we’ve delved into four in our Floodwater Recharge Memos.

Benefits Diagrams_Page_5

The first is identifying the beneficiaries and the potential magnitude of those benefits. As can be seen in the flow chart above, there many more potential beneficiaries than just the local groundwater users. Some of these benefits require forecast informed reservoir operations (FIRO) to realize those gains through reduced flood control space, increased water supply storage and greater summertime hydropower output. Flood-MAR programs can provide the needed margin of error to lower the risk from FIRO.

FloodMAR Poster - Financing

The second is finding the funding mechanisms to compensate growers or to build dedicated recharge basins. We prepared a list of potential financing mechanisms linked to the potential beneficiaries. (This list grew out of another study that we prepared for the Delta Protection Commission on feasible options for beneficiary-pays financing.)

FloodMAR Poster Incentives

The third is determining what type of market incentive transactions mechanisms would work best at attracting the most preferred operations and acreage. I have explored the issues of establishing unusual new markets for a couple of decades, including for water rights transfer and air quality permit trading. It is not a simple case of “declaring markets exist” and then walking away. Managing institutions have important roles in setting up, running and funding any market, and most particularly for those that manage what were “public goods” that individuals and firms were able to use for free. The table above lists the most important considerations in establishing those markets.

The fourth assessing what type of infrastructure investment will produce the most cost-effective recharge. Construction costs (which we evaluated) is one aspect, and impacts on agricultural operations and financial feasibility are other considerations. The chart at the top summarizes the results from comparing several case studies. These will vary by situation, but remarkably, these options appear to cost substantially less than any surface storage projects currently being proposed.

The final institutional issue to be addressed, but not the least important, is determining the extent of rights over floodwaters and aquifers. California state law and regulations are just beginning to grapple with these issues. Certain areas are beginning to assert protection of their existing rights. This issue probably represents the single biggest impediment to these programs before attracting growers to participate.

All of these issues can be addressed in a range of pilot programs which use different variables to test which are likely to be more successful. Scientists and engineers can use these pilots to test for the impacts of different types of water diversion and application. Statistical regression analysis can provide us much of what we know without having to understand the hydrological dynamics. Legal rights can be assessed by providing temporary permits that might be modified as we learn more from the pilots.

Is it time to move forward with local pilot programs? Do we know enough that we can demonstrate the likely benefits? What other aspects should we explore before moving to widespread adoption and implementation?