Close
About
FAQ
Home
Collections
Login
USC Login
Register
0
Selected
Invert selection
Deselect all
Deselect all
Click here to refresh results
Click here to refresh results
USC
/
Digital Library
/
University of Southern California Dissertations and Theses
/
Intelligent signal processing for oilfield waterflood management
(USC Thesis Other)
Intelligent signal processing for oilfield waterflood management
PDF
Download
Share
Open document
Flip pages
Contact Us
Contact Us
Copy asset link
Request this asset
Transcript (if available)
Content
INTELLIGENT SIGNAL PROCESSING FOR OILFIELD WATERFLOOD MANAGEMENT by Feilong Liu A Dissertation Presented to the FACULTY OF THE GRADUATE SCHOOL UNIVERSITY OF SOUTHERN CALIFORNIA In Partial Fulfillment of the Requirements for the Degree DOCTOR OF PHILOSOPHY (ELECTRICAL ENGINEERING) May 2008 Copyright 2008 Feilong Liu ii Dedication To Brycen, Chuping and my parents iii Acknowledgments I would like to express my deepest gratitude and thanks to my advisor, Professor Jerry M. Mendel, for his invaluable guidance, inspiration, support, encouragement and patience through all these years in all aspects of my research and career. He has taught me the way to do scientific research, the way to be an authority in researching fields, and the way to be the member or leader of a team. I would also like to express my great gratitude to Dr. Iraj Ershaghi for his consistent support and suggestions on this research; Drs. Iraj Ershaghi and B. Keith Jenkins for serving on my qualifying exam and dissertation committee; and Drs. Antonio Ortega and Shrikanth S. Narayanan for serving on my qualifying exam committee. My gratitude also goes to Dave Tuk, Jim Brink, Dave Rossi, John Houghton, Tai Nguyen and Vega Sankur for their very helpful recommendations on Chevron job, and Steve Giffoni for providing the summer internship opportunity and job offer. I also would like to thank my officemates, teammates (e.g., Amir Mohammad Nejad), basketball buddies, soccer buddies and many friends at USC. With them, the life in US becomes very interesting and memorable. Finally, I would like to appreciate my beloved wife, chuping, my funny son, Brycen, and my beloved parents for their unconditional support. With their companion, understanding, encouragement and support, my life in US becomes much cozy, blissful and fruitful. iv Table of Contents Dedication ii Acknowledgements iii List of Tables vi List of Figures viii Abstract xiii 1 Introduction 1 1.1 Motivation and objective 1 1.2 Thesis outline 4 2 Forecasting Injector-producer Relationships from Production and Injection Rates Using an Extended Kalman Filter 5 2.1 Introduction 5 2.2 Reservoir model 8 2.2.1 Scale function 11 2.2.2 Injector-producer subsystem model 12 2.2.3 Injector-producer relationship 15 2.3 Extended Kalman Filter 19 2.3.1 State Variable Model 19 2.3.2 Extended Kalman Filter 23 2.3.3 Production forecaster using SVM 26 2.4 Simulations 28 2.4.1 Synthetic Data 28 2.4.2 CMG Reservoir Simulation Data 42 2.4.3 Real Data 44 2.5 Conclusions 49 3 Inferring Regional Impacts for Multiple Injectors and Multiple Producer, Using a Pseudo-virtual Reservoir Approach 50 3.1 Introduction 50 3.2 Pseudo-virtual reservoir approach 51 3.2.1 Reservoir model 51 3.2.2 Virtual reservoir model 54 3.2.3 Pseudo-virtual reservoir model 60 3.2.4 Estimator using Extended Kalman Filter 63 3.3 Simulations 68 v , 3.3.1 Procedure for choosing the values of , 69 ,, (, ) jl j l fr k , , jl jl a b 3.3.2 Synthetic data generation 70 3.3.3 Monte-carlo simulation 71 3.4 Conclusions 100 4 Background Knowledge for Type-1 and Type-2 Fuzzy Sets 101 4.1 Type-1 Fuzzy Sets 101 4.2 Type-2 Fuzzy Sets 103 5 Encoding Words into Interval Type-2 Fuzzy Sets Using an Interval Approach 113 5.1 Introduction 113 5.2 Interval Approach 118 5.2.1 Interval approach: data part 120 5.2.2 Interval approach: fuzzy set (FS) part 128 5.2.3 Observations 141 5.3 Codebook Examples 143 5.4 Conclusions 150 6 Conclusions and Future Works 151 6.1 Conclusions 151 6.2 Future Works 153 6.2.1 Perform field test 153 6.2.2 Construct a Decision Support System Using CWW 155 6.2.3 Some future works about interval approach for type-2 fuzzistics 158 Appendix A 159 Appendix B 164 Appendix C 169 References 174 List of Tables Table 2.1 The chosen parameters for Case 1. 35 Table 2.2 The chosen parameters for Case 2 before 1000 k = . 39 Table 2.3 The chosen parameters for Case 2 after 1000 k = . 39 Table 3.1 Summarization of the settings for the test cases. 68 Table 3.2 The parameters for Case 1. 73 Table 3.3 The parameters for case 2. 75 Table 3.4 The parameters for Case 3 when 1, ,167 k = … . 78 Table 3.5 The parameters for Case 3 when 168, ,500 k = … . 78 Table 3.6 The parameters for Case 4 when 1, ,167 k = … . 80 Table 3.7 The parameters for Case 4 when 168, ,500 k = … . 80 Table 3.8 The parameters for Case 5. 83 Table 3.9 The parameters for Case 6. 86 Table 3.10 The parameters for Case 7 when 1, ,167 k = … . 90 Table 3.11 The parameters for Case 7 when 168, ,500 k = … . 90 Table 3.12 The parameters for Case 8. 93 Table 3.13 The parameters for Case 9 when 1, ,167 k = … . 97 Table 3.14 The parameters for Case 9 when 168, , 500 k = … . 97 Table 5.1 Tolerance factor k for a number of collected data ( m ′ ), a proportion of the data (1 α − ), and a confidence level 1 γ − [43]. 123 Table 5.2 Mean and standard deviation for interior and shoulder T1 MFs [16]. 130 vi vii ] Table 5.3 Transformations of the uniformly distributed data interval () () [, i i ab into the parameters () i MF a and () i MF b of a T1 FS [16]. 131 Table 5.4 Remaining data intervals and their end-point statistics for m data intervals. 144 Table 5.5 FOU data for all words (based on data intervals)—the Codebook. 145 m * List of Figures Figure 1.1 Block diagram of our research program 2 Figure 2.1 Reservoir model for N injectors and a single producer 8 Figure 2.2 Two examples of the impulse response between an injector and a producer. 12 Figure 2.3 An example of channel impulse response and 14 ( ) j h t ( ) j h n Figure 2.4 An example of (a) step change of injection rate; and (b) consequent change of production rate. 15 Figure 2.5 The diagram for implementing EKF estimator and production forecaster. 32 Figure 2.6 An example of generated measured production rates for Case 1. 36 Figure 2.7 The mean of estimated ( ) j IPR k Λ for Case 1. 36 Figure 2.8 The standard deviations of estimated ( ) j IPR k Λ for Case 1. 37 Figure 2.9 The mean of forecasted production rates. 37 ( ) e m k Figure 2.10 The standard deviations ( ) e k σ of forecasted production rates. 38 Figure 2.11 An example of generated measured production rates for Case 2. 40 Figure 2.12 The mean of estimated ( ) j IPR k Λ for Case 2. 40 Figure 2.13 The standard deviations of the estimated ( ) j IPR k Λ for case 2. 41 Figure 2.14 The mean of forecasted production rates. 41 ( ) e m k Figure 2.15 The standard deviations ( ) e k σ of forecasted production rates. 42 Figure 2.16 The reservoir model for simulation data generation. 43 viii Figure 2.17 The estimated ( ) j IPR k Λ for case 3. 44 Figure 2.18 Estimated IPRs between producer W5 12-1A and its surrounding injectors W5 12-1B and W5 11-1W. 46 Figure 2.19 Estimated IPRs between producer W5 12-1A and its surrounding injectors W5 12-1B and W5 11-1W and a distant injector W5 11-2W. 46 Figure 2.20 Estimated IPRs between producer W5 12-1A and its surrounding injectors W5 12-1B and W5 11-1W and a distant injector W5 11-1WA. 47 Figure 2.21 Estimated IPRs between producer W5 12-1A and its surrounding injectors W5 12-1B and W5 11-1W and a distant injector W5 9-2W. 47 Figure 2.22 Estimated IPRs between producer W5 10-3 and its surrounding injectors W5 9-3W W5 11-3WBS and W5 10-2WAS. 48 Figure 3.1 An example of a region with four injectors and two producers. 51 Figure 3.2 Reservoir model for N injectors and M producers. 53 Figure 3.3 An example of virtual reservoir with four injectors and two producers. 55 Figure 3.4 An example of virtual impulse response . 57 () v j h t Figure 3.5 Virtual Reservoir Model for N injectors and M producers. 59 Figure 3.6 An example for the virtual impulse response and the scaled ( ) v j h t pseudo-impulse response (, ) ( pv pv pv jj j ) frk h t . 61 Figure 3.7 Pseudo-virtual reservoir model for N injectors and M producers. 61 Figure 3.8 The mean of estimated ˆ (| ) R j IPk k for Case 1. 74 Figure 3.9 The standard deviation of estimated ˆ ( | ) R j IPk k for Case 1. 74 Figure 3.10 The mean of ˆ (| ) R j IPk k for Case 2. 76 Figure 3.11 The standard deviation of ˆ ( | ) R j IPk k for Case 2. 76 ix Figure 3.12 The mean of ˆ (| ) R j IPk k for Case 3. 79 Figure 3.13 The standard deviation of ˆ ( | ) R j IPk k for Case 3. 79 Figure 3.14 The mean of ˆ (| ) R j IPk k for Case 4. 81 Figure 3.15 The standard deviation of ˆ ( | ) R j IPk k for Case 4. 82 Figure 3.16 The mean of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 5. 84 Figure 3.17 The standard deviation of ˆ ( | ) R j IPk k using injection rates from only close injectors for Case 5. 84 Figure 3.18 The mean of ˆ (| ) R j IPk k using injection rates from all injectors for Case 5. 85 Figure 3.19 The standard deviation of ˆ ( | ) R j IPk k using injection rates from all injectors for Case 5. 85 Figure 3.20 The mean of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 6. 87 Figure 3.21 The standard deviation of ˆ ( | ) R j IPk k using injection rates from only close injectors for Case 6. 87 Figure 3.22 The mean of ˆ (| ) R j IPk k using injection rates from all injectors for Case 6. 88 Figure 3.23 The standard deviation of ˆ ( | ) R j IPk k using injection rates from all injectors for Case 6. 88 Figure 3.24 The mean of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 7. 91 Figure 3.25 The standard deviation of ˆ ( | ) R j IPk k using injection rates from only close injectors for Case 7. 91 x Figure 3.26 The mean of ˆ (| ) R j IPk k using injection rates from all injectors for Case 7. 92 Figure 3.27 The standard deviation of ˆ ( | ) R j IPk k using injection rates from all injectors for Case 7. 92 Figure 3.28 The mean of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 8. 94 Figure 3.29 The standard deviation of ˆ ( | ) R j IPk k using injection rates from only close injectors for Case 8. 95 Figure 3.30 The mean of ˆ (| ) R j IPk k using injection rates from all injectors for Case 8. 95 Figure 3.31 The standard deviation of ˆ (|) R j IPk k i ZI ∧ using injection rates from all injectors for Case 8. 96 Figure 3.32 The mean of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 9. 98 Figure 3.33 The standard deviation of ˆ ( | ) R j IPk k using injection rates from only close injectors for Case 9. 98 Figure 3.34 The mean of ˆ (| ) R j IPk k using injection rates from all injectors for Case 9. 99 Figure 3.35 The standard deviation of ˆ ( | ) R j IPk k using injection rates from all injectors for Case 9. 99 Figure 4.1 A type-1 triangular membership function and one of its alpha-cuts 102 Figure 4.2 Membership function of ( ) A α 102 Figure 4.3 An example of a type-2 membership function 104 Figure 4.4 An example of the vertical slice for a particular type-2 membership function 107 xi Figure 4.5 An example of embedded type-2 and type-1 fuzzy set 110 Figure 5.1 The Perceptual Computer (Per-C) for CWW using IT2 FSs. 114 Figure 5.2 Data Part of the IA approach. Note that the data statistics, S 1 ,...,S m , and interval statistics, m and , feed into the FS Part of IA, in Fig. 5. The extra heavy lines and blocks denote the flow l m r of processing once the data are collected. 119 Figure 5.3 Left-shoulder, right-shoulder and interior FOUs, all of whose LMFs and UMFs are piecewise linear. 119 Figure 5.4 An example of two overlapping data intervals for the same word. 125 Figure 5.5 FS Part of the IA approach. The extra heavy lines and blocks denote the flow of processing once the data are collected. 129 Figure 5.6 Admissible region for an interior FOU that is based on (5.14) and (5.15). 133 Figure 5.7 Admissible region for an interior FOU that is based on (5.16) and (5.17). 134 Figure 5.8 FOU decision regions. 135 Figure 5.9 An example of the union of T1 triangle MFs. The heavy lines are the lower and upper MFs for the interior FOU. 139 Figure 5.10 An example of the union of T1 left-shoulder MFs. The heavy lines are the lower and upper MFs for the left-shoulder FOU. 139 Figure 5.11 An example of the union of T1 right-shoulder MFs. The heavy lines are the lower and upper MFs for the right-shoulder FOU. 139 Figure 5.12 FOUs for all 32 words. Start at the top row and proceed downwards scanning from left-to-right. 146 Figure 5.13 FOUs for all 32 words when Stage 4 of data pre-processing is omitted. Start at the top row and proceed downwards scanning from left-to-right. 147 Figure 5.14 FOUs for three sub-vocabularies. 148 xii xiii Abstract This thesis addresses two problems about water-flood management: (1) infer reservoir heterogeneity using measured injection and production rates, and (2) construct a decision support system to optimize oil production using computing with words (CWW) and the inferred reservoir heterogeneity. To infer reservoir heterogeneity, we first present an adaptive method using an Extended Kalman Filter (EKF) for the case of multiple injectors and a single producer, and then present a pseudo-virtual reservoir method for the case of multiple injectors and multiple producers, respectively. In the EKF method, a very simple parametric model, one with two parameters per injector, is used so that if a producer depends upon N injectors our model contains exactly 2N parameters; and the EKF is used to adaptively estimate the 2N parameters, from which the injector-producer relationship (IPR) between each injector and a producer is then estimated. In the pseudo-virtual reservoir method, a virtual reservoir model is used to model the reservoir, and a pseudo- virtual reservoir model is used to estimate the regional impact from each injector, so that the problem for this case reduces to the problem for the case of multiple injectors and a single producer. The study for decision support system using CWW focuses on the encoder component (called type-2 fuzzistics) of a perceptual computer (Per-C), a specific architecture for CWW using interval type-2 fuzzy sets, and our work so far has been to transform linguistic perceptions, words, into interval type-2 fuzzy sets (IT2 FS) that activate a CWW engine. We have proposed a new and simple approach, called the xiv Interval Approach (IA), to type-2 fuzzistics, one that captures the strong points of both the person membership function and interval end-points approaches. It collects interval end-point data from subjects, does not require subjects to be knowledgeable about fuzzy sets, has a straightforward mapping from data to footprint of uncertainty (FOU), does not require an a priori assumption about whether or not a FOU is symmetric or non- symmetric, and leads to an IT2 FS word model that reduces to a T1 FS word model automatically if all subjects provide the same intervals. 1 Chapter 1 Introduction 1.1 Motivation and Objective As the oil demand increases globally, and the numbers of newly discovered significant oilfields decrease, the requirement to produce oil more efficiently and economically becomes more and more urgent. Primary recovery using natural producing mechanisms, i.e., liquid and rock expansion and solution gas drive, leave behind 80% or more of the original oil in place (OOIP). Consequently, the secondary recovery becomes more significant than ever to recover billions of barrels of additional oil from oil reservoirs. Various secondary recovery methods, e.g., water-flooding, steam-flooding, CO 2 flooding, hydrocarbon gas injection and chemical injection etc, have been proposed. Water-flooding is one of the most widely used methods because water is easy to obtain and is inexpensive [40]. Unfortunately, the combined primary and current secondary recovery still leave about two-thirds of the OOIP behind. This stimulates the demand to bring more technologies into water-flood management. Water-flood management is a broad topic involving a range of activities, some of which are performed only once or twice during the life of the water-flood, e.g., water- flood design and facilities/well planning. Other water-flood activities are performed periodically based on analyses of recurring measurement data, e.g., setting water injection flow rates in the injection wells. The success of water-flood management depends on an understanding of reservoir heterogeneity. The more we know about the reservoir, the greater the chance water-flooding can succeed. One way to infer the reservoir heterogeneity is to determine some expression (often either mathematically or empirically derived) of the inter-relationship between injectors and corresponding affected producers using measured production and injection rates, which are the most abundant data available in oilfield. In this thesis, this inter-relationship is referred as the “injector-producer relationship (IPR)” for the case of multiple injectors and a single producer, and the “regional impact” for the case of multiple injectors and multiple producers. By knowing the IPR and regional impact, we might be able to tell something about the directional sweep efficiency of a given pattern, obtain a better understanding of zonal conformance and reservoir heterogeneity, and make better decisions on waterflood management. Production 2 Figure 1.1: Block diagram of our research program. As depicted in Fig. 1.1, the objectives of this thesis is to address two problems about water-flood management: (1) infer reservoir heterogeneity using measured injection and Infer Reservoir Heterogeneity Reservoir Heterogeneity Decision Support System Using Computing With Words and Decisions Injection rates Expert Rules 3 production rates, and (2) construct a decision support system to optimize oil production using computing with words (CWW) and the inferred reservoir heterogeneity. To infer reservoir heterogeneity, we first present an adaptive method using an Extended Kalman Filter (EKF) for the cases of multiple injectors and a single producer, and then present a pseudo-virtual reservoir approach for the case of multiple injectors and multiple producers, respectively. In the EKF method, a very simple parametric model, one with two parameters per injector, is used so that if a producer depends upon N injectors our model contains exactly 2N parameters; and the EKF is used to adaptively estimate the 2N parameters, from which the injector-producer relationship (IPR) between each injector and a producer is then estimated. In the pseudo-virtual reservoir method, a virtual reservoir model is used to model the reservoir, and a pseudo- virtual reservoir model is used to estimate the regional impact from each injector, so that the problem for the latter reduces to the problem for the former. The study for decision support system using CWW focuses on the encoder component of a perceptual computer (Per-C), called type-2 fuzzistics. The Per-C is described in Chapter 5 (see Fig. 5.1) and is a specific architecture for CWW that uses interval type-2 fuzzy sets. Our work as far has been to transform linguistic perceptions, words, into interval type-2 fuzzy sets (IT2 FS) that activate a CWW engine. We have proposed a new and simple approach to type-2 fuzzistics, called the Interval Approach (IA), one that captures the strong points of both the person membership function and interval end-points approaches. It collects interval end-point data from subjects, does not require subjects to be knowledgeable about fuzzy sets, has a straightforward 4 mapping from data to footprint of uncertainty (FOU), does not require an a priori assumption about whether or not a FOU is symmetric or non-symmetric, and leads to an IT2 FS word model that reduces to a T1 FS word model automatically if all subjects provide the same intervals. 1.2 Thesis Outline The rest of this thesis is organized as follows. In Chapter 2, we begin with reservoir model, scale function and injector-producer model, define the injector-producer relation, construct the state variable model, and explain the EKF approach. Tests using synthetic data, CMG simulation data and real data are given to provide us some preliminary results on this EKF approach. Chapter 3 begins with a virtual reservoir model, defines the regional impact, explains the similarity between this virtual reservoir model and the reservoir model in Chapter 2, and introduces a pseudo- virtual reservoir model to estimate regional impact. Tests using synthetic data confirm the feasibility of this virtual reservoir approach. Chapter 4 reviews some required and concise background knowledge about type-1 and type-2 fuzzy logic systems that are needed in the subsequent chapters. Chapter 5 briefly outlines the decision support system using computing with words, after which a new and simple approach, called the Interval Approach (IA), to type-2 fuzzistics is presented. Finally, in Chapter 6, we summarize our work and propose some future work. 5 Chapter 2 Forecasting Injector-Producer Relationship from Production and Injection Rates Using Extended Kalman Filter 2.1 Introduction Water-flood management is a broad topic involving a range of activities, some of which are performed only once or twice during the life of the water-flood, e.g., water- flood design and facilities/well planning. Other water-flood activities are performed periodically based on analyses of recurring measurement data, e.g., setting water injection flow rates in the injection wells. The success of water-flood management depends on our understanding of reservoir heterogeneity. The more we know about the reservoir, the greater chance the water-flooding can succeed. One way to infer the reservoir heterogeneity is to determine some expression (often either mathematically or empirically derived) of the inter-relationship between injectors and corresponding affected producers using the production and injection rates, the most abundant data available in the oilfield. Usually, this inter-relationship is referred to as the “injector- producer relationship (IPR)” for the case of multiple injectors and a single producer. By knowing the relative IPRs, we might be able to tell something about the directional sweep efficiency of a given pattern or even directional fracture, leading to a better understanding of zonal conformance and reservoir heterogeneity, and maximizing the oil production by changing the injection rates from different injectors. Many methods have been previously used to infer IPRs using only injection and production rates [1], [2], [6], [35]-[37], [39], [50], among which, the most recent works are from Albertoni and Lake [1], and Yousef et al [50]. In their works, the reservoir is considered to be a system, which can be represented as a continuous impulse response that converts the input signal (i.e., injection rates) into an output signal (i.e., production rates). Albertoni and Lake model the reservoir by a resistive model (RM), which is characterized by a collection of weights, whereas Yousef et al. model the reservoir by a capacitance model (CM), which is characterized by two parameters, λ and τ , that quantify the connectivity and the dissipation between wells, respectively. The IPRs (i.e., the interwell connectivities) are quantitatively evaluated by the model parameters, i.e., weights for RM, and λ for CM. The parameters in these models are estimated using the same approach, multiple linear regression (MLR); however, there are some weak points of both approaches: (1) In the process of estimating the IPRs, the parameters in the models and the IPRs are assumed stationary, i.e., unchanged over the window during which the data are analyzed so that when the IPRs change the analysis needs to be repeated for the new situation, and this may not be practical because the reservoir is dynamic and it may be very difficult to recognize when a change has occurred; (2) The CM model is very complex, and although it is characterized by two parameters λ and τ , to use it, the primary production and bottom hole pressure impact also need to be determined. 6 7 To overcome these difficulties, we propose a new approach to determine the IPRs in which [17]: 1) Learning from expert knowledge and the shape of a diffusivity filter, an exact two-parameter auto-regressive (AR) model is used to model the impulse response between a single injector and a single producer, and the area under the impulse response is then used as a measure of the IPR. If a producer depends on N injectors, our model will contain exactly 2N parameters; 2) Instead of assuming stationarity, the IPRs are considered to be non-stationary, i.e., IPRs may change constantly due to changing the bottom hole pressure, work-over or earthquake, but we have no idea about when the changes will happen, and how often they will occur; 3) The Extended Kalman Filter (EKF) is used to forecast and track the non- stationary IPR parameters because, as shown below, the state variable model (SVM) for our IPR model is nonlinear. Our forecasts of IPRs are nonlinear and adaptive functions of the measured data, where the nature of the nonlinearities is established by the mathematics of the EKF, and adaptation automatically permits the forecasters to track the changing conditions. In this Chapter, we focus on the multiple injectors and single producer situation; the work for the case of multiple injectors and multiple producers will be discussed in Chapter 3. 2.2 Reservoir Model Our “reservoir model” is a high-level system-oriented model and is not a petroleum engineering model. As such, it views a reservoir as an interconnection of injectors to a producer, and interprets this as a communication system in which the injectors are transmitting messages to the producer, but the messages are distorted by the channel between each injector and the producer. One could call this a synthetic reservoir model, but for short, we call it a reservoir model. 1 ( ) it 8 Figure 2.1: Reservoir model for N injectors and a single producer. The reservoir model for N injectors and a single producer is depicted in Fig. 2.1, in which each injector-producer pair is considered as an independent subsystem; hence, for a reservoir with N injectors and a single producer, the reservoir is represented by N 1 ( ) ht f(r 1 , k 1 ) ∑ ,1 ( ) m it 1 ( ) nt 2 ( ) i t ∑ ,2 ( ) m it 2 ( ) nt ( ) N i t ∑ , ( ) mN it ( ) N n t f(r 2 , k 2 ) f(r N , k N ) ( ) N h t 1 ( ) c p t ( ) p n t 2 ( ) ht Σ 2 () c p t () c N () p t ( ) m p t Σ p t independent subsystems, where each subsystem is modeled as a continuous-time communication channel. The notations in this reservoir model are: 1) , and ( () j it () j nt , ( ) mj i t 1, , j N = … ) are the actual injection rates that flow into the reservoir, the corresponding measurement noises for measuring injection rates, and the measured injection rates, respectively; 2) (, ) j j f rk ( ) are linear or nonlinear scalar functions of 1, , j = …N j r and , which relate to the connectivities between injectors and the producer, where j k j r and are distance and permeability between injector j and the producer, respectively. Heuristically, the greater the value of j k (, ) j j f rk the better is the connectivity between the injectors and producer; 3) ( ( ) j ht 1, , j N = … ) is the channel impulse response for subsystem j that is used to model the continuous-time communication channel between injector j and the producer. Each impulse response will be modeled as described in (2.4), and is characterized by two parameters; 4) () p t , and () p nt ( ) m p t are the actual production rate, the corresponding measurement noise for measuring production rate, and the measured production rate, respectively. (Note that production rate means the total production rate, i.e., the sum of oil, gas and water production rates.); 5) () c j p t is the channel production rate produced by injector j, which represents the amount of production rate in ( ) p t caused by injector j; 9 6) , , ( ) Ps ( ) c j P s ( ) j I s and are Laplace transforms of ( ) j H s () p t , () c j p t , , and , respectively; and () j it () j h t 7) , , ( ) Pz ( ) c j P z ( ) j I z and are z-transforms of ( ) j H z ( ) p n , ( ) c j p n , and , respectively. ( ) j i n ( ) j h n It will be seen later that: 1) The noise-free signals , () j it () c j p t ( 1, , j N = … ) and () p t are used for constructing reservoir model, whereas the measured signals ( , ( ) mj i t 1, , j N = … ) and () m p t are used for estimating the parameters of the reservoir model during our processing, because the noise-free signals , () j it () c j p t ( ) and 1, , j = …N ( ) p t are not directly available; 2) The Laplace transforms , , ( ) Ps ( ) c j P s () j I s and are used to describe the reservoir model, whereas the z-transforms , , ( ) j H s ( ) Pz ( ) c j P z () j I z and are used for estimating the parameters of the reservoir model during our processing, because only discrete measurement data are available for such processing. ( ) j H z From Fig. 2.1, we see that 1 () ( ) N c j j ptp = = ∑ t (2.1) and () ( , ) ( ) ( ) c jjjj j ptfrk it ht ⎡ ⎤ =∗ ⎣ ⎦ (2.2) 10 where * is the convolution operation. Taking the Laplace transform of both sides of (2.1) and (2.2), it follows that 11 j (2.3) 11 () () ( , ) ( ) ( ) NN c jjjj jj Ps P s f r k H s I s == == ∑∑ 2.2.1 Scale function (, ) j j f rk Our view is that ( , ) j j f rk is a linear or nonlinear scalar function of and . To date, we have no explicit formula for that. Even if we had an explicit formula for j r j k (, ) j j f rk , it would still be impossible for us to use it because we don’t know the permeability of the channel between the injector and producer; however, we do have some useful heuristic knowledge about ( , ) j j f rk , namely, 1) (, ) j j f rk is a monotonically decreasing function of ; j r 2) (, ) j j f rk is a monotonically increasing function of ; and, j k 3) The larger ( , ) j j f rk is, the better is the connectivity from the injector to producer; the shorter is the delay time (i.e., the larger is ), and the larger is the impact from the injectors (i.e., the larger is j a j IPR ). This knowledge will be used by us for our experimental designs when we have to choose the parameters for different reservoir models. Also, as we will see in Section 2.2.4, the ( , ) j j f rk will be absorbed into j γ when we estimate j IPR . 2.2.2 Injector-Producer Subsystem Model Each injector-producer pair is modeled as a continuous-time impulse response that converts the injection rates into production rate. Learning from expert knowledge (i.e., Chevron Engineers) and the shape of a diffusivity filter, this continuous-time impulse response has been assumed to be the continuous-time uni-modal function depicted in Fig. 2.2, which can be characterized as ( ) at ht bte − = . (2.4) The Laplace transform of , has two poles at ( ) ht () Hs a − , i.e., 2 () () b Hs sa = + (2.5) Figure 2.2: Two examples of the impulse response between an injector and a producer. 12 Because only sampled data (i.e., measurements of injection and production rates) are available for processing, the continuous-time impulse response is sampled, and a discrete impulse response is used to approximate it. Two ways to model a discrete impulse response that has a uni-modal shape are: (1) Use an auto-regressive (AR) model that has two equal poles, e.g., for a single injector and single producer, such an AR model is 13 ) 2 (1) 2 () ( 1) ( pkpk pk αα γ +− + − =ik (2.6) where () p k and are the production and injection rates, respectively, and, and ( ) ik aT e α − = b T γ α = are the two parameters that characterize the model. (2) Use a moving-average (MA) model, e.g., for a single injector and single producer, the MA model is 1 () ( ) l j j pk ik β = j = − ∑ (2.7) where () p k and are the production and injection rates, respectively, ( ) ik j β , ( 1, , j l = … ) are the weights, and l is the length of the impulse response. Usually, to better approximate the continuous-time impulse response, the value of l is much greater than 2. The AR model has the following advantages over the MA model: (1) It needs fewer parameters to describe the discrete impulse response than the MA model, i.e., for N injectors, the AR model only needs 2N parameters, whereas the MA model needs 1 N i i l = ∑ parameters; and (2) It is difficult to determine the lengths of the MA model, i.e., the values of . Consequently, our preference for this project is to use the AR model, which has the following z-transform formulation: 1 ,, N l l … 1 12 () (1 ) j j j z Hz z γ α − − = − (2.8) where j aT j e α − = and jjj b T γ α = are the two parameters that characterize this discretized channel model. An example of such a and is depicted in Fig. 2.3, where the solid curve represents and the dots represent the non-zero values of . Note that (2.8) is derived directly from (2.4). ( ) j h t ( ) j h n ( ) j h t ( ) j h n t ( ) j h t Figure 2.3: An example of channel impulse response and ( ) j h t ( ) j h n (the solid dots represent the non-zero samples). 14 2.2.3 Injector-Producer Relationship (IPR) As shown in Fig. 2.4, when the j th injection rate has a step change ( ) jj I ut t ∆ − at time , it produces a change in j t () c j p t , ( ) c j p t ∆ . Because the Laplace transform of () jj I ut t ∆ − is 1 j st j I e s − ∆ , it follows from (2.3) that: 11 1 1 () () ( , ) ( ) ( ) ( , ) () j NN N st c jjjjj jjjj jj j Ps P s f r k H s I s f r k H s I e s − == = ∆= ∆ = ∆ = ∆ ∑∑ ∑ (2.9) j I ∆ j I ∆ 15 Figure 2.4: An example of (a) step change of injection rate; and (b) consequent change of production rate. t () c j p t ( ) c j p t ∆ t j t (a) j t (b) Taking the inverse Laplace transform of (2.9), the change of channel production rate produced by injector j, ( ) c j p t ∆ , can be expressed as: 0 () ( , ) ( ) ( ) ( , ) ( ) ( ) t c jjj j jj jjj jj p t f rk Iut t h t I f rk u t h t d τ τ τ ⎡⎤ ∆= ∆ − ∗ =∆ − − ⎣⎦ ∫ (2.10) When , the steady-state value of t →∞ ( ) c j p t ∆ , ( ) c j p ∆ ∞ , is 0 0 () ( , ) ( ) ( ) ( , ) ( ) ( , ) ( ) j c jjjj jj jj j j t jj j j p Ifr k u t h d If r k h d If r k h t dt τ ττ ττ ∞ ∞ ∞ ∆∞=∆ − ∞− =∆ ∞ − =∆ ∫ ∫ ∫ (2.11) The second line of (2.11) has made use of the fact that ( ) 0 j ut τ − = when j t τ < and ( ) j ut 1 τ− = when j t τ > . The third line of (2.11) has made use of the transformation of variables t τ =∞ − . Definition 1: The impact from injector to producer ( ) j it ( ) p t , j IP , is defined as 0 () (, ) () c j jjj j p j IPfrkh I ∞ ∆∞ ≡= ∆ ∫ tdt (2.12) 16 This formula shows that a unit step change of the injection rate at an injector, i.e., , can cause a step change of steady-state production rate that equals 1 j I ∆= 0 (, ) () jj j f rk h tdt ∞ ∫ . Obviously, is the area under the continuous-time impulse response . 0 () j htdt ∞ ∫ ( ) j ht In order to compute j IP , we need ; however, instead of using , our method for computing () j ht ( ) j ht j IP is to use the sampled value of , , and to approximate ( ) j ht () j hn j IP in (2.12) as 0 0 1 2 (, ) () ( , ) ( ) ( , ) ( ) | (, ) (1 ) jjj j jj j n jj j z jj j j IPfrk htdt Tf r k h n Tf r k H z fr k T γ α ∞ ∞ = = = ≈ = = − ∫ ∑ (2.13) The fact that 1 0 () ( )| j j n hn H z ∞ z = = = ∑ follows from the definition of 0 () ( ) n jj n Hz h nz ∞ − = = ∑ . Because the sample period T is normally the same for all injectors, we omit it and evaluate the impact using the relative injector-producer relationship. 17 Definition 2: The injector-producer relationship for the j th injector, j IPR , is defined as 1 2 (, ) (, ) ( )| (1 ) jj j j jj j IP f r k IPR f r k H z T jjz γ α = =≈ = − (2.14) In order to estimate j IPR from measured data, we need to estimate j α , j γ and (, ) j j f rk ; however, ( , ) j j f rk , which can be a linear or nonlinear scalar function of and , is unknown. Conceptually, since j r j k j γ and (, ) j j f rk appear as a product, it is impossible to resolve each individually; hence, we re-express j IPR in (2.14) as 2 (1 ) jj j j IP IPR T γ α ′ =≈ − (2.15) where (,) jj j j f rk γ γ ′ = . (2.15) shows that each IPR is characterized by two parameters j α and j γ ′ . The process to estimate j IPR becomes: 1) Estimate j α and j γ ′ , and obtain ˆ j α and ˆ j γ ′ ( ) 1, , j = …N ; and, 2) Estimate j IPR by j IPR Λ , where 2 ˆ ˆ (1 ) j j j IPR γ α Λ ′ = − (2.16) This is done by the EKF. 18 2.3 The Extended Kalman Filter Using the available sampled measurements of injection and production rates, our goals are: 1) Estimate the channel impulse response parameters j γ ′ and j α () 1, , j N = … ; 2) Compute j IPR Λ using ˆ j γ ′ and ˆ j α ( ) 1, , j N = … ; and, 3) Forecast the future production rate. 2.3.1 State Variable Model (SVM) The first step to using an EKF is to construct a SVM [19]. To illustrate how to do this, consider a simple case with three injectors and a single producer (it is easy to generalize the SVM from three injectors to N injectors.), for which the reservoir model in (2.3) is: 12 3 11 1 2 2 2 3 3 3 1 3 12 1 () () ( ) () ( ) () () ( ) () () ( ) () ( ) ( ) (1 ) cc c j j j j Pz Pz Pz Pz f r H zI z f r H zI z f r H zI z z Iz z γ α − − = =+ + =+ + ′ = − ∑ (2.17) where ( () ( ) () () c jjj Pz f r H zI z = j 3 1,2, j = ) are z-transforms for the respective production rates from injectors 1, 2 and 3, respectively. Separate from to obtain ( ): ( ) c j P z ( ) Pz 1, 2, 3 j = 19 11 12 1 2 2 () () (1 ) 1 2 c jj j jj j Pz z z j Izz z γγ αα −− zα − −− ′′ == −− + (2.18) (2.18) can be expressed as: () (2.19) 122 1 12 () () c jj j j j zzPz zI αα γ −− − ′ −+ = z Multiply on both sides of (2.19), and determine the inverse z-transform of the resulting equation, to see that (2.19) can be expressed in the time-domain as: z (2.20) () 2 , , (1) 2 () ( 1) () ( ) ( ) ( ) ( ) c j cc c jjj jj jj jmj j jmj p pk pk pk i k ik nk ik n k αα γ γ γ ′ +− + − = ′=− ′=+ where can be seen from Fig. 2.1, and , () () ( ) jmj j ik i k n k = − () ( ) c j jj p nk n k γ ′ =− . This is a second-order finite-difference equation that is described by two state variables (1 c j pk ) − and ( ) c j p k . In addition, the unknown parameters, i α and i γ are also treated as state variables so that they can be estimated by an EKF, and their state equation models are (1) () () (1) () () j j jj jj kkn kknk α γ αα γγ ′ += + ⎫k ⎪ ⎬ ⎪ ′′ += + ⎭ (2.21) 20 where and are discrete-time zero mean white random noises. Eqs. (2.20) and (2.21) together are described by a ( ) j n k α ( ) j n k γ ′ 4 1 × state vector j x , where ( ), ( ), ( 1), ( ) cc jj j j j k k pk pk αγ ′ ′ ⎡ ⎤ =− ⎣ ⎦ x (2.22) Using standard techniques [19], the SVM for c j p is ( 1, 2, 3 j = ): (2.23) 1 1 2 2 4 3 2 14 1 3 4 2, () (1) (1) () (1) (1) (1) () ( (1) () 2() () () () (1) (1) () () ( 1) 0 0 j j j j j j j jj c jj jj j j c j j jmj c j xk k xk xk k xk kxk xk p k xkx k xkx k xk pk xki k pk α γ ⎡⎤ + ⎡⎤ + ⎢⎥ ⎡⎤ ⎢⎥⎢⎥ ⎢⎥ + + ⎢⎥⎢⎥ ⎢⎥ += = = + ⎢⎥⎢⎥ ⎢⎥ + ⎢⎥⎢⎥ ⎢⎥ − ⎢⎥⎢⎥ ⎢⎥ + + ⎣⎦ ⎣⎦ ⎢⎥ + ⎣⎦ += x xn () 0 1 ( 1) j k ⎧ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎨ ⎪ ⎪ ⎪ ⎪ ⎪ + ⎪ ⎩ x )k ) 3 where and are the respective production rate and measured injection rate for injector j ( ), and, ( 1 c j pk + , ( ) mj i k 1, 2, j = ( ) [ ( ), ( ), 0, ( )] c ij j j p kn k n k n k αγ ′ = x n (2.24) is a additive zero-mean white noise for the state equations. 4 1 × 21 Let [ ] 12 3 (1) ( 1), (1), ( 1) kcol k k k +≡ + + + xx x x ; then, the complete SVM for all three injector subsystems is: 11 11 12 12 14 13 2 11 14 11 13 12 ,1 14 21 22 23 24 31 32 33 34 () (1) () (1) () (1) 2() () () () () ( (1) (1) (1) (1) (1) (1) (1) (1) (1) (1) m xk xk xk xk xk xk xkx k xkx k x ki k xk xk xk k xk xk xk xk xk xk + ⎡⎤ ⎢⎥ + ⎢⎥ ⎢⎥ + ⎢⎥ −+ + ⎢⎥ ⎢⎥ + ⎢⎥ + ⎢⎥ += = ⎢⎥ + ⎢⎥ ⎢⎥ + ⎢⎥ + ⎢⎥ ⎢⎥ + ⎢⎥ + ⎢⎥ ⎢⎥ + ⎣⎦ x 21 22 24 2 21 24 21 23 22 ,2 31 32 34 2 31 34 31 33 32 ,3 12 3 ) () () () () 2() () () () () () () () () 2() () () () () () (1) ( 1) (1) ( 1) m m cc c m xk xk k xk xkx k xkx k x ki k xk xk xk x kxk x kx k xki k p k pk pk pk ⎡⎤ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ + ⎢⎥ ⎢⎥ −+ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ −+ ⎣⎦ + = ++ ++ ++ x n () (1) 0 0 0 1 0 0 0 1 0 0 0 1 ( 1) ( 1) p p nk knk ⎧ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎨ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ + ⎪ ⎪ =+ ⎩ x++ (2.25) where is the measured production rate, ( 1) m pk + 12 3 () [ ( ), (), ( )] kcol k k k = xx x x nnnn and are additive zero-mean white noises for the state and measurement equations with covariance matrix and variance ( 1 p nk +) k Q 1 k r + , respectively, where (2.26) 11 2 2 3 3 12 , , 0, , , , 0, , , , 0, cc pp nn n n n n n n n diag r r r r r r r r r αγ α γ α γ ′′ ′ ⎡⎤ = ⎢⎥ ⎣⎦ k Q 3 c p 22 23 ) Because this SVM is nonlinear, we can’t use a Kalman filter to estimate its twelve states; Instead, we use the EKF to estimate them. Note that after all twelve states are estimated, we can extract the estimates of the six injector-producer model parameters from them, namely ˆ(1| 1 j kk α + + , ˆ(1| 1 j kk ) γ ′ + + ( 1,2, j 3 = ), after which we can compute ( ˆ (1| 1 j IPR k k + +) 3 1,2, j = ) as 2 ˆ(1| 1) ˆ (1| 1) ˆ (1 ( 1| 1)) j j j kk IPR k k kk γ α ′ + + ++ = −+ + (2.27) In this way, we see that the estimates of ˆ j IPR are functions of time. 2.3.2 EKF approach The EKF has been widely used [19], [5] as the standard technique for recursive nonlinear estimation. It provides a first-order approximation to optimal nonlinear mean- square state estimation for the following nonlinear discrete-time system: [ ] [] (1) (), () (1) ( 1), 1 (1), 1,2, y kf kk k yk h k k n k k ⎧ += + ⎪ ⎨ ⎪ += + + + + = ⎩ … x xx n x (2.28) where and are additive zero-mean white noises for the state and measurement equations with covariance matrices and variance , respectively. () k x n ( ) y nk ( ) k Q ( ) rk 24 k + x ˆ(1|) k k The EKF operates in two stages, predictor and corrector. The predictor computes the predicted value of , (1) + x k + x ˆ(1| 1) kk + + x ) , and the corrector computes the filtered value of , . (1) In the EKF, the state equations are linearized about , and the measurement equation is linearized about ˆ( | ) kk x ˆ(1| k k + x , i.e., [ ] [ ] [ ] [][][ ] ˆˆ ˆ ( 1) ( |), ( |), ( ) ( |) ( ) ˆˆ ˆ (1) ( 1|), 1 ( 1|), 1 ( 1) (1|) (1 y kkkk kkkkxkk k yk h k kk kkk k kk nk ⎧ +≈ + − + ⎪ ⎨ ⎪ +≈ + + + + + + − + + + ⎩ xx x xfx x x n xx xx F H ) (2.29) where [ ] ˆˆ (1|) (|), kk kkk += xfx , [ ] (), ( ) kk k =∂ ∂ x F fx x and [ ] (), ( ) hkk k =∂ ∂ x H xx . Note that x F is a Jacobian matrix, and 4 4 N N × x H is a 14N × Jacobian matrix for the SVM of the reservoir model. Example 1: For a single injector, using (2.23), it follows that: 2 14 11 13 ,1 11 11 10 0 01 0 00 0 2 ( ) 2 () () () () 2 ( ) m 0 0 1 xkxkxk i k x k xk ⎡⎤ ⎢⎥ ⎢⎥ = ⎢⎥ ⎢⎥ −− ⎣⎦ x F (2.30) and [ ] 0 0 0 1 = x H ; (2.31) and, for three injectors, using (2.25), it follows that: 1 2 3 ⎡ ⎤ ⎢ ⎥ ⎢ ⎥ = ⎢ ⎥ ⎢ ⎥ ⎣ ⎦ 00 00 00 x A F A A (2.32) and [ ] 0 0 0 1 0 0 0 1 0 0 0 1 = x H . (2.33) where (2.34) 4 2 ,1 1 13 10 0 0 01 0 0 00 0 1 2() () () 2 ( ) 2() () i mi i i ii xk ik x k x k xkx k ⎡⎤ ⎢⎥ ⎢⎥ ⎢⎥ = ⎢⎥ − ⎢⎥ − ⎢⎥ ⎣⎦ i A for ( ), and 1,2,3 i = 0 is a 4-by-4 zero matrix. The EKF is summarized as follows: 1. Initialize the EKF with , , and . ˆ(0 | 0) x (0|0) P k Q k r 2. Predictor ( ): 0,1, k = … 25 [ ] ˆˆ (1|) (|), kk f kkk += xx (2.35) ˆˆ (1|) ((|),)(|)((|),) k kk kkk kk kkk ′ += + xx PFx PFx Q (2.36) 3. Corrector ( ): 0,1, k = … [ ] { } ˆˆ ˆ (1| 1) ( 1|) ( 1) (1) ( 1|), 1 kk kk k yk h kkk + + = +++ +− + + xx K x (2.37) 1 ˆ (1|) (( 1|)) (1) ˆˆ (( 1| )) ( 1| ) (( 1| )) k kk kk k kk kk kk + ′ + + += ′ ++ + + x xx PHx K HxP Hx r ) (2.38) (2.39) ˆ (1| 1) [ (1) ((1|))](1| kk k k k k k ++ = − + + + x PIKHx P 2.3.3 Production forecaster using SVM (Production Forecaster) Using the SVM in (2.28), the future regional production rate can be forecasted using zero state and measurement noises and scheduled injection rates , i.e., , ( ) sj i k [ ] [] (1) (), ( 1) ( 1), 1 , , 1, kf kk yk h k k k n n ⎧ += ⎪ ⎨ ⎪ += + + = + ⎩ … xx x (2.40) Example 2: For a single injector, using (2.40) and (2.23), it follows that the future channel production rate ( 1 c j pk ) + is forecasted as 26 () 1 2 4 2 14 1 3 2, () () (1) () 2() () () () () () ( 1) 0 0 0 1 ( 1) j j jj jj j j jsj c jj xk xk kxk x kx k x kx k xki k pk k ⎧ ⎡ ⎤ ⎪ ⎢ ⎥ ⎪ ⎢ ⎥ ⎪ ⎢ ⎥ += ⎪ ⎢ ⎥ ⎪ ⎢ ⎥ − ⎪ ⎢ ⎥ ⎨ ⎢ ⎥ + ⎪ ⎣ ⎦ ⎪ ⎪ ⎪ ⎪ += + ⎪ ⎩ x x (2.41) and, for three injectors, using (2.40) and (2.25), it follows that the future regional production rate ˆ ( ) m p t is forecasted as 11 12 14 2 11 14 11 13 12 ,1 21 22 24 2 21 24 21 23 22 ,2 31 32 34 2 31 34 31 33 32 ,3 () () () 2() () () () () () () () (1) () 2() () () () () () () () () 2() () () () () () s s s xk xk xk x kx k x kx k x k i k xk xk k xk x kx k x kx k x k i k xk xk xk x kx k x kx k x k i k ⎡ ⎢ ⎢ ⎢ ⎢ −+ ⎢ ⎢ ⎢ ⎢ ⎢ += ⎢ ⎢ −+ ⎢ ⎢ ⎢ −+ ⎣ x ( 12 3 ( 1) ( 1) ( 1) ( 1) 0 0 0 1 0 0 0 1 0 0 0 1 ( 1) cc c m p k pk pk pk k ⎧ ⎤ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎥ ⎨ ⎥ ⎪ ⎥ ⎪ ⎥ ⎪ ⎢⎥ ⎪ ⎢⎥ ⎪ ⎢⎥ ⎪ ⎢⎥ ⎦ ⎪ ⎪ ⎪ += + + + + + ⎪ ⎪ =+ ⎩ x) (2.42) The forecasting is implemented as follows: 1. Initialize in SVM using . 1 ( ) k x 1 1 ˆ(| ) kk x 27 2. Set , and let be the length of window for forecasting. 1 n k = fw L 28 ) 3. Forecast using (2.41) for single injector, and (2.42) for three injectors. ˆ( 1 c j pk + 4. Set ; if 1 nn = + 1 fw nk L < + , go to Step 3; otherwise, go to Step 5. 5. Stop. 2.4 Simulations Our approach was tested using synthetic data, CMG simulation data and real data. The tests using synthetic data used Monte-Carlo simulations, the tests using CMG simulation data used the data generated by petroleum commercial simulator, and the test using real data used data from Section 5 of Chevron’s Lost Hills Field. The reason for using synthetic data is that we can validate our approach because we know what “truth” is. We can easily alter data sets to reflect plausible reservoir geology changes, so that we can know if our approach can adapt to such changes. Tests using reservoir simulation data are performed to confirm whether or not our estimated results match the models used for generating reservoir simulation data. Tests using real data are performed to see if our results match some expert knowledge. 2.4.1 Synthetic data This section begins with the procedure for choosing the values of the parameters (, ) j j f rk , in reservoir model, introduces the procedure for synthetic data generation and Monte-Carlo simulation, and finally summarizes the tests. , j a b j A. Procedure for choosing the values of (, ) j j f rk , . , jj a b Recall from our third heuristic-knowledge for (, ) j j f rk in Section 2.2.1, the larger (, ) j j f rk is, the better is the connectivity from the injector to producer; the shorter is the delay time (i.e., the larger is ), and the larger is the impact from the injectors (i.e., the larger is j a j IPR ). Our procedure for choosing the values of (, ) j j f rk , is: , j a b j 1) Specify the value of ( , ) j j f rk in [ ] 0, 1 ; 2) Choose the value of [ ] 1,1000 j a ∈ and [ ] 0, 1 j IPR ∈ using following rules: Rule 1: If ( , ) j j f rk is large, then and j a j IPR are large. Rule 2: If (, ) j j f rk is moderate, then and j a j IPR are moderate. Rule 3: If ( , ) j j f rk is small, then and j a j IPR are small. 3) Determine using j b j aT j e α − = , jjj b T γ α = and (2.8) as () () () 11 22 2 ()| ( )| 1 1 (, )1 jj jj jjz jz j aT aT j jj j aT aT jj Hz Hz b T T ee IPR fr k e e T γ α αα == −− −− == = − − = − T (2.43) where T is a small sample period, which corresponds to high sampling rate so that the generated discrete signal can be considered as a continuous signal. 29 B. Synthetic data generation Synthetic data was generated as follows: (Note: the procedure is explained for the case of three injectors and two producers, and can be extended to any other situation.) Step 1: Generate the noise-free injection rates ( ( ) h j i n 1, 2, 3 j = ) using the injection rate model described in Appendix A. Step 2: To generate the noise-free production rates ( ) h p n , the SVM for multiple injectors and a single producer was implemented with zero-noise by specifying ( , ) j j f rk , ( , jj a b 1, 2, j 3 = ) using the procedure described in Section 2.4.1.A and a small 0.01 T = , which corresponds to a high sampling rate (recall that j aT j e α − = and jjj b T γ α = ). Step 3: The noise-free injection rates ( ( ) j i k 1, 2, 3 j = ) and noise-free production rates ( ) p k were obtained by down-sampling ( ) and ( ) h j i n 1, 2, j = 3 ( ) h p n by 20. This corresponds to a low sampling rate. Step 4: The measured injection and production rates were obtained by adding measurement noises with () j nk () () 2 10 2 () 10log 20 () j j Ei k SNR En k ⎡⎤ ⎢ ⎥ = = ⎢⎥ ⎣⎦ dB to ( ), and with ( ) j i k 1,2, j = 3 ( ) p n k () () 2 10 2 () 10log 20 () p Epk SNR En k ⎡⎤ ⎡⎤ ⎢⎥ ⎣ ⎦ = = ⎢⎥ ⎡⎤ ⎢⎥ ⎢⎥ ⎣⎦ ⎣⎦ dB to ( ) p k , i.e., 30 , () () () () () () mj j j mp ik ik nk pkpk nk =+ ⎧ ⎪ ⎨ ⎪ =+ ⎩ (2.44) C. Procedure for implementing EKF estimator and production forecaster Recall that our goals are to estimate the channel impulse response parameters j γ ′ and ˆ j α ; compute ( 1, , j N = … ) ˆ j IPR using j γ ′ and ˆ j α ( ) 1, , j N = … ; and forecast the production rate ˆ() p t ; and these works have been done either separately or simultaneously. An example with four time zones ( 01 kt k ≤ ≤ , , 11 fw kt k L ≤≤ + 11 2 fw kt k L ≤≤ + and 1 2 fw tk L ≥+ .) is depicted in Fig. 2.5, in which, 1) For , the measured injection and production rates, ,( 0 kt k ≤≤ 1 ) , ( ) mj it m p t , flow into the EKF estimator; and are obtained; ˆ ˆ (), ( ), ( ) jj j ttIPR γα ′ t 2) For 11 fw kt k L <≤ + , where fw L is the length of window for production forecasting, the estimated flow into both the EKF estimator and production forecaster as the initialization such that (a) are continuously updated using the new incoming injection and production rates, ,( 11 ˆ ˆ (), ( ), ( ) jj j kkIPRk γα ′ 1 t ) ˆ ˆ (), ( ), ( ) jj j ttIPR γα ′ , ( ) mj it m p t , and (b) ˆ( ) p t is forecasted from time to 1 k 1 fw k L + using the scheduled injection rates . , ( ) sj i t 3) For 1 1 2 fw kL t k L fw + <≤ + , the estimated , 11 ˆ (),( jfwj kL kL γ α ′ ++) fw 31 32 w 1 ˆ () jf IPR k L + flow into both the EKF estimator and production forecaster as the initialization such that (a) are continuously updated using the new incoming injection and production rates, it,( ˆ ˆ (), ( ), () jj j ttIPR γα ′ t ) , ( ) mj m p t , and (b) the ˆ( ) p t is forecasted after time using the scheduled injection rates . 2 k , ( ) sj i t 4) For 1 2 fw tk L >+ , do the same as in step 3 except the initialization becomes , 11 ˆ (2 ), (2 ) fw jfwj kL k L γ α ′ ++ 1 ˆ (2 ) jfw IPR k L + . 0 k Figure 2.5: The diagram for implementing EKF estimator and production forecaster. EKF Estimator EKF Estimator Production Forecaster () EKF Estimator Production Forecaster , ( ) m j it m p t t ˆ ˆ (), ( ), ( ) jj j ttIPR γα ′ 1 k t ˆ( ) p t ˆ( ) p t ˆ ˆ (), ( ), ( ) jj j ttIPRt γα ′ ˆ ˆ (), ( ), ( ) jj j ttIPRt γα ′ , ( ) sj it it , ( ) sj m () p t m () p t , ( ) mj , ( ) mj it 1 fw kL + it 1 2 fw kL + D. Monte-Carlo simulations Simulations were implemented as follows: Step 1: Set , number of realizations 1 i = 100 NR = , and the number of samples in each realization 500 M = for Case 1 and 2000 M = for Case 2. Step 2: For the realization ( th i 1, 2,3 j = , 1, , kM = … ): 1) Generate and , ( ) mj i k ( ) m p k , using the four-step procedure described in Section 2.4.1.B. 2) Implement the EKF using the just generated and , ( ) mj i k ( ) m p k , and obtain ˆ(| ) j kk γ ′ and ˆ(| ) j kk α ; 3) Compute the ˆ (| ) j i IPR k k ⎡ ⎣ ⎤ ⎦ using (2.27). 4) Forecasting the production rate ˆ( ) p k , using the procedure described in Section 2.4.1.C. 5) Compute the forecasting error [ ] () i ek using [ ] ˆ () () ( ) m i ek p k p k = − (2.45) Step 3: If , go to Step 4; otherwise, set i NR ≥ 1 ii = + and go to Step 2. Step 4: Compute the mean and standard deviation of ˆ (|) j i IPR k k ⎡ ⎤ ⎣ ⎦ using 1 1 ˆ () ( | ) NR jj i i mk IPR k k NR = ⎡ ⎤ = ⎣ ⎦ ∑ (2.46) 33 and ( ) 2 1 1 ˆ () ( | ) ( ) NR j j i i kIPRkkm NR σ = ⎡⎤ = ⎣⎦ ∑ j k− , (2.47) respectively. Step 5: Compute the mean and standard deviation of [ ] () i ek using [] 1 1 () () NR e i i mk ek NR = = ∑ (2.48) and []( ) 2 1 1 () () () NR e i i kekm NR σ = = − ∑e k, (2.49) respectively. Step 6: Stop. E. Tests on synthetic data Many tests were performed, and two of them are presented below. The first is for constant IPRs, and the second is for non-constant IPRs. Case 1: Three injectors and a single producer with constant IPRs For this case, the reservoir is static with constant IPRs. The parameters for this case are summarized in Table 1. Observe from the values of j IPR that injector 2 has the 34 largest impact on the producer, injector 1 has about 75% less impact on the producer than injector 2, and injector 3 has the smallest impact on the producer, and this is about 95% less than injector 2. Table 2.1: Parameters for Case 1. j (, ) jj f rk j a j b j α j γ j IPR 1 0.5 200 120 0.1353 0.1624 0.1086 2 1 400 2400 0.0183 0.4396 0.4561 3 0.25 10 0.1 0.9048 0.0009 0.025 An example of measured production rates is depicted in Fig. 2.6. The test results are depicted in Figs. 2.7 and 2.8 for estimates of ( ) j IPR k ( 1, 2, 3 j = ), and in Figs. 2.9 and 2.10 for the production forecast error with 1 125 k = and 90 fw L = , respectively. Note that the original values of ( ) j IPR k are provided in the legend of Fig.2.7. Observe that: 1) From Fig. 2.7, our approach can correctly estimate the relative IPRs, i.e., the largest IPR Λ is obtained from the most influential injector, and the smallest IPR Λ is obtained from least influential injector. 2) The and () e mk ( ) e k σ , which range mostly from -50 to 50, and 30 to 50, respectively, are much smaller than the generated measured production rates, which is around 1200. This confirms that our approach can successfully forecast future production rates. 35 0 50 100 150 200 250 300 350 400 450 500 500 1000 1500 2000 k The measured production rates y(k) Figure 2.6: An example of generated measured production rates for Case 1. 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The mean of estimated IPRs m 1 (k) 0.11 m 2 (k) 0.46 m 3 (k) 0.025 Figure 2.7: The mean of estimated ( ) j IPR k Λ for Case 1. 36 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The standard deviations of estimated IPRs σ 1 (k) σ 2 (k) σ 3 (k) Figure 2.8: The standard deviations of estimated ( ) j IPR k Λ for Case 1. 100 150 200 250 300 350 400 450 500 -150 -100 -50 0 50 100 150 200 k The mean of forecasted production rates error m e (k) Figure 2.9: The mean of forecasted production rates. ( ) e m k 37 100 150 200 250 300 350 400 450 500 25 30 35 40 45 50 55 k The standard deviation of forecasted production rates error σ e (k) Figure 2.10: The standard deviations ( ) e k σ of forecasted production rates. Case 2: Three injectors and a single producer with non-constant IPRs Unlike Case 1, for which the IPRs were constant for all measurements, in this case, we assume that something such as a work-over has happened so that the reservoir model is different before and after 1000 k = . The parameters for this case are summarized in Tables 2 and 3. Observe that, before 1000 k = , injectors 1 and 2 have the largest and smallest impacts on the producer, respectively; and, after 1000 k = , injector 2 becomes the most influential injector, whereas the influence from injector 1 reduces to around 45% of that from injector 2. An example of measured production rates is depicted in Fig. 2.11. Test results are depicted in Figs. 2.12 and 2.13 for estimates of ( ) j IPR k ( 1, 2, j 3 = ), and in Figs. 2.14 38 and 2.15 for the forecast error of production rates with and 1 500 k = 90 fw L = , respectively. Note that the original values of ( ) j IPR k are provided in the legend of Fig. 2.12. Observe that: 1) Our approach can correctly estimate the relative IPRs, i.e., the largest IPR is obtained from the most influential injector, and the smallest IPR is obtained from least influential injector. Λ Λ 2) Around , there is a change of the mean of the estimated 1000 k = ( ) j IPR k Λ , which shows that our approach can correctly and quickly detect and track the changes of the IPRs. 3) The and () e mk ( ) e k σ are much smaller than the generated measured production rates, which confirms that our approach can successfully forecast the future production rates. 4) Around , the and 1000 k = ( ) e m k ( ) e k σ have a big change, which shows that the forecasting errors relate to the changes of the IPRs directly. Table 2.2: Parameters for Case 2 before 1000 k = . j (, ) jj f rk j a j b j α j γ j IPR 1 1 200 300 0.1353 0.4060 0.5430 2 0.5 56 5.8 0.5703 0.0335 0.0907 3 0.8 100 41.8 0.3679 0.1539 0.3082 Table 2.3: Parameters for Case 2 after 1000 k = . j (, ) jj f rk j a j b j α j γ j IPR 1 0.7 200 120 0.1353 0.1624 0.1520 2 1 400 2400 0.0183 0.4396 0.4561 3 0.4 51 3.3 0.6 0.02 0.05 39 0 200 400 600 800 1000 1200 1400 1600 1800 2000 500 1000 1500 2000 2500 3000 k The measured production rates y(k) Figure 2.11: An example of generated measured production rates for Case 2. 0 200 400 600 800 1000 1200 1400 1600 1800 2000 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The mean of estimated IPRs m 1 (k) 0.5430 0.1520 m 2 (k) 0.0907 0.4561 m 3 (k) 0.3082 0.05 Figure 2.12: The mean of estimated ( ) j IPR k Λ for Case 2. 40 0 200 400 600 800 1000 1200 1400 1600 1800 2000 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The standard deviations of estimated IPRs σ 1 (k) σ 2 (k) σ 3 (k) Figure 2.13: The standard deviations of the estimated ( ) j IPR k Λ for case 2. 500 1000 1500 2000 -500 0 500 1000 k The mean of forecasted production rates error m e (k) Figure 2.14: The mean of forecasted production rates. ( ) error m k 41 500 1000 1500 2000 0 50 100 150 200 250 300 k The standard deviation of forecasted production rates error σ e (k) Figure 2.15: The standard deviations ( ) error k σ of forecasted production rates. 2.4.2 Simulation data The simulation data were generated by Amir Mohammad Nejad using a commercial simulator. A simple five spot water injection pattern was adopted; and, the permeability, thickness and distance in a real oilfield were used to construct the models for generating simulation data. Different settings of permeability, thickness and distance were used, and data were generated for a broad range of models. Many tests were performed, and only one of them is presented herein. 42 Injector 3 Injector 2 Injector 4 Figure 2.16: The reservoir model for simulation data generation. Case 3: Five-Spot Static Reservoir The reservoir model used for this case is depicted in Fig. 2.16. Observe that the permeability between injector 2 and producer has the highest value, then the permeability between injector 1 and producer, the permeability between injector 4 and producer, and the permeability between injector 3 and producer has the smallest value. Consequently, injector 2 has the largest impact on the producer; then, injectors 1, 4, 3 have smaller impacts on the producer in that order. Test results are depicted in Fig. 2.17 for the estimated IPRs. Observe that the estimated ( ) j IPR k Λ match the model, i.e., the largest ( ) j IPR k Λ is obtained from the most 43 influential injector, and the smallest ( ) j IPR k Λ is obtained from the least influential injector. 1900/02/08 1905/08/01 1911/01/22 1916/07/14 1922/01/04 1927/06/27 1932/12/17 1938/06/09 1943/11/30 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Time The estimated IPRs Injector 1 Injector 2 Injector 3 Injector 4 Figure 2.17: The estimated ( ) j IPR k Λ for Case 3. 2.4.2 Real data A real dataset was provided by Chevron. It is weekly data from 30 injectors and 35 producers in Section 5 at Lost Hills. Recall that the objective of tests using real data is to see if our results match the expert knowledge. Two expert knowledges about the Lost Hills oilfield are: (1) Lost Hills is an oilfield with low permeability, so that injectors distant from the producer should have little impact upon it; and (2) there is a natural 44 fracture around North-East 55 degree at Chevron Lost Hill field. Many tests were performed, and two of them are presented herein. Case 4: Reservoir model with two close injectors and three distant injectors In this case, we use a two-step procedure to test our approach: Step 1: We chose the producer W5 12- 1A, and its surrounding injectors W5 12-1B and W5 11-1W, for which 1 163.5532 r = and 2 138.0518 r = , and obtained the estimated IPRs, as shown in Fig. 2.18; and, Step 2: We add a distant injector (i.e., one that is far away from the producer such that one wouldn’t anticipate it would have any influence on the producer.) at one time, and obtained the estimated IPRs depicted in Figs. 2.19 to 2.21, respectively, for three injectors W5 11-2W, W5 11-1WA and W5 9-2W with , 556.7273 and 1393.3. 3 639.6760 r = From Figs. 2.18-2.21, observe that (1) the IPRs are dynamic, i.e., they change with time; (2) distant injectors impact the producer very little, which matches the first expert knowledge about Lost Hills that Lost Hills is an oilfield with low permeability, so that injectors distant from the producer should have little impact upon it; and, (3) the estimated IPRs of the surrounding injectors change very little when a distant injector is included. 45 0 50 100 150 200 250 300 350 400 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 k The estimated IPRs W5 12-1B W5 11-1W Figure 2.18: Estimated IPRs between producer W5 12-1A and its surrounding injectors W5 12-1B and W5 11-1W. 0 50 100 150 200 250 300 350 400 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 k The estimated IPRs W5 12-1B W5 11-1W W5 11-2W Figure 2.19: Estimated IPRs between producer W5 12-1A and its surrounding injectors W5 12-1B and W5 11-1W and a distant injector W5 11-2W. 46 0 50 100 150 200 250 300 350 400 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 k The estimated IPRs W5 12-1B W5 11-1W W5 11-1WA Figure 2.20: Estimated IPRs between producer W5 12-1A and its surrounding injectors W5 12-1B and W5 11-1W and a distant injector W5 11-1WA. 0 50 100 150 200 250 300 350 400 0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 k The estimated IPRs W5 12-1B W5 11-1W W5 9- 2W Figure 2.21: Estimated IPRs between producer W5 12-1A and its surrounding injectors W5 12-1B and W5 11-1W and a distant injector W5 9-2W. 47 Case 5: Reservoir with three close injectors. For this case, we chose the producer W5 10-3 and its surrounding injectors W5 9-3W, W5 11-3WBS and W5 10-2WAS, for which 1 333.5987 r = , and , and obtain estimated IPRs depicted in Fig. 2.22. 2 306.7820 r = 3 251.8346 r = 0 50 100 150 200 250 300 350 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 k The estimated weights W5 9-3W W5 11-3WBS W5 10-2WAS Figure 2.22: Estimated IPRs between producer W5 10-3 and its surrounding injectors W5 9-3W W5 11-3WBS and W5 10-2WAS. Observe that: (1) There is a change at about 150 k = , which suggests something has happened, e.g., a work-over. This shows that our approach can quickly detect changes of the IPRs; (2) it appears that W5 9-3W has the most impact on the producer; however, from the distance, we know that this injector is the farthest one from the producer. An explanation for the second observation is there may be a fracture connection between 48 49 injector W5 9-3W and producer W5 10-3, which matches the second expert knowledge about Lost Hills, i.e., that there is a natural fracture around North-East 55 degree at Chevron Lost Hill field. 2.5 Conclusions This chapter presents an adaptive method, using an extended Kalman filter (EKF), to forecast the injector-producer relationships (IPRs) between multiple injectors and a single producer based on measured production and injection rates. Our approach has been tested on synthetic data, reservoir simulation data and real data. Test results on synthetic data and reservoir simulation data demonstrate the feasibility of the EKF method, and test results on the real data match expert knowledge about the IPRs between injectors and a producer. All results confirm that this EKF method can provide a good way to infer and track the IPRs so as to provide better insight about the IPRs. Our approach may also be able to provide information about the directional sweep efficiency of a given pattern, a directional fracture, and may lead to a better understanding of zonal conformance and reservoir heterogeneity, all of which will assist Chevron engineers to maximize the oil production by changing the injection rates from different injectors. Chapter 3 Inferring Regional Impact for Multiple Producers and Multiple Injectors, Using a Pseudo-virtual Reservoir Approach 3.1 Introduction Current optimal decisions for water-flood management are made on a pattern basis, where a pattern generally consists of a group of injectors and producers. A region can be a single pattern or multiple patterns. The change of injection rate on a single injector may simultaneously have positive or negative impacts on multiple producers. Consequently, the impact of a single injector on a region may be more important than its impact on a single producer. If a regional impact is available, better decisions could be made to increase or decrease the injection rates for a particular injector in the region, such that the total oil production from the region is increased. In this thesis, our focus is on inferring the regional impact from each injector. By doing so, we will obtain a collection of tools that are able to provide flexibility for decision making under different scales, e.g., a region with a single producer, a collection of producers, a pattern, or a collection of patterns. 50 3.2 Pseudo-virtual reservoir approach This section begins with the reservoir model, derives the virtual reservoir model and finally introduces the pseudo-virtual reservoir model. Consider the simple example of a region depicted in Fig. 3.1, in which there are four injectors that surround two producers. Instead of determining the impact on a particular producer from each injector, which can be done by using a straight-forward extension of our results in Chapter 2 from one producer to more than one producer, our objective is to determine the regional impact from each injector on the region that encircles the two producers. The approach proposed in this report builds upon our previous work in Chapter 2 for multiple injectors and a single producer. Injector Producer Figure 3.1: An example of a region with four injectors and two producers. 3.2.1 Reservoir model For generality, we assume that there are N injectors and M producers in a region. The reservoir model for such a region is depicted in Fig. 3.2, in which the N injectors 51 and each producer pair is considered as an independent subsystem; hence, for a reservoir with N injectors and M producers, the reservoir is represented by M independent subsystems, where each subsystem is modeled as the reservoir model depicted in Fig. 2.1. The notations for this reservoir model are: 1) , and ( () j it () j nt , ( ) mj i t 1, , j N = … ) are the actual injection rates that flow into the reservoir, their corresponding measurement noises, and the measured injection rates, respectively; 2) , , (, ) jl jl frk ( ; 1, , jN = … 1, , l M = … ) are linear or nonlinear scalar functions of , j l r and , which relate to the connectivities between injectors and producers, where , jl k , j l r and are the distance and permeability of the reservoir channel between injector j and producer l, respectively; , jl k 3) ( ; , () jl h t 1, , jN = … 1, , l M = … ) is the channel impulse response for the reservoir channel between injector j and producer l; it is modeled by a continuous-time uni-modal function, similar to (2.4), i.e., , , , () j l a t jl jl ht bte − = ; 4) () l p t , and () l p n t , () ml p t are the actual production rate, their corresponding measurement noise, and the measured production rate for producer l, respectively; 5) , ( ) l Ps () j I s and are the Laplace transforms of , ( ) jl H s ( ) l p t , , and , respectively; and () j i t , () jl h t 6) , ( ) l Pz () j I z and are the z-transforms of , ( ) jl H z ( ) l p n , , and , respectively. () j in , ( ) jl h n 52 1 ( ) it Figure 3.2: Reservoir model for N injectors and M producers. 1 ( ) it ∑ ,1 ( ) m it 1 ( ) nt 2 ( ) i t ∑ ,2 ( ) m it 2 ( ) nt ( ) N i t ∑ , ( ) mN it ( ) N n t 1, ( ) M ht f(r 1,M , k 1,M ) f(r 2,M , k 2,M ) f(r N, M , k N, M ) , ( ) NM ht 2, ( ) M ht Σ ( ) M p t Σ ( ) M p nt , ( ) mM p t 2, () c M 1, ( ) c M pt p t , ( ) c NM p t ∑ 1,1 ( ) ht f(r 1,1 , k 1,1 ) ,1 ( ) m it 1 ( ) nt 2 ( ) i t ∑ ,2 ( ) m it 2 ( ) n t ( ) N i t ∑ , ( ) mN it ( ) N n t f(r 2,1 , k 2,1 ) f(r N,1 , k N,1 ) ,1 ( ) N ht 2,1 ( ) ht 1,1 () c p t 1 () p n t 2,1 () c p t Σ 1 ( ) p t ,1 ( ) m p t Σ ,1 ( ) c N p t 53 3.2.2 Virtual reservoir model To begin, we introduce some new definitions. Definition 1: A regional producer is a pseudo-producer that comprises all producers in a region, with actual regional production rate, () R p t , and measured regional production rates, () R m p t , given by: 1 () () M R l l ptp = = ∑ t (3.1) () , 11 1 1 () () () () () () () ( ) R ll MM M M R R mml l p l p p ll l l ptpt ptnt pt nt ptn == = = == + = + = + ∑∑ ∑ ∑ t (3.2) where is the measurement noise for the actual regional production rate. 1 () ( ) R l M p p l nt n t = = ∑ Definition 2: A virtual channel is a virtual communication channel between an injector and the regional producer, which converts the injection rates into regional production rates. An example of regional producer and virtual channels for the reservoir depicted in Fig. 3.1 is shown in Fig. 3.3, in which the two producers are replaced by a single regional producer, and the dashed line from each injector to the regional producer denotes a virtual channel. We refer to this reservoir as a virtual reservoir. 54 Injector Producer Regional Producer Figure 3.3: An example of virtual reservoir with four injectors and two producers. Taking the Laplace- and z- transforms of both sides of (3.1), it follows that . (3.3) 1 () ( ) M R l l Ps Ps = = ∑ and . (3.4) 1 () ( ) M R l l Pz Pz = = ∑ It will be seen later that: (1) () j it () 1, , j N = … and ( ) l p t ( ) 1, , l M = … are used for constructing the virtual reservoir model; and and , ( ) mj i t , ( ) ml p t are used for estimating the parameters of the virtual reservoir model during our processing, because the noise-free signals and () j it ( ) l p t are not directly available; 55 (2) , , ( ) l Ps () R P s ( ) j I s and are used for constructing the virtual reservoir model, and , , () , ( ) jl H s ( ) l Pz () R P z j I z and are used for estimating the parameters of the virtual reservoir model during our processing, because only discrete measurement data are available for such processing. , ( ) jl H z Using (2.3), but for the l th producer, can be expressed as ( ) l Ps (3.5) ,, , 1 () ( , ) ( ) ( ) N ljljljl j Ps f r k H s I s = = ∑ j j jl Substituting (3.5) into (3.3), it follows that (3.6) ,, , 111 ,, , 11 1 () () ( , ) ( ) ( ) ( , ) ( ) ( ) ( ) ( ) MMN R ljljljl llj NM jl jl jl j jl N v jj j Ps Ps f r k H sI s fr k H s I s HsI s === == = == ⎧⎫ = ⎨⎬ ⎩⎭ = ∑∑∑ ∑∑ ∑ where is the Laplace transform of the virtual channel between injector j and the regional producer, i.e., ( ) v j H s (3.7) ,, , 1 () ( , ) ( ) M v jjljl l Hs f r k H s = = ∑ 56 Taking the inverse Laplace transform of both sides of (3.7), it follows that the virtual impulse response can be expressed as ( ) v j h t (3.8) ,, , 1 () ( , ) ( ) M v jjljl l ht f r k h t = = ∑ jl . which shows that the impulse response of a virtual channel is the weighted sum of M actual channel impulse responses. Although each is characterized by a two parameter uni-modal function [see (2.4)], will not be a two parameter uni-modal function. Instead, it is a sum of M AR model and has 2 , () jl h t ( ) v j h t M parameters. An example of for two producers (i.e., ( ) v j h t 2 M = ) is shown in Fig. 3.4, where the dotted and dashed curves represent the scaled impulse responses from injector j to producers 1 and 2, respectively, and the solid curve (which is the sum of these two curves) represents the virtual impulse response . ( ) v j h t ( ) h • Figure 3.4: An example of virtual impulse response . ( ) v j h t ,1 ,1 ,1 (, ) () jj j frk h t t ,2 ,2 ,2 (, ) ( jj j ) frk h t ( ) v j h t 57 Let () v j p t be the virtual channel production rate due to injector j and the j th virtual channel, its Laplace transform is given by . (3.9) () () ( ) vv jj j Ps H sI s = Substituting (3.9) into (3.6), it follows that the actual reservoir model can be expressed as: (3.10) 11 () () () ( ) NN Rv v jj jj Ps P s H sI s == == ∑∑ j Observe from (3.10) that, by collecting multiple producers in a region as a regional producer, and by introducing virtual channels with virtual impulse responses, the actual reservoir model can be reformulated as a virtual reservoir model, as depicted in Fig. 3.5. Comparing Figs. 3.5 and 2.1, observe that the virtual reservoir model is actually a reservoir model for multiple injectors and a single producer with ( , ) 1 jj fr k = and ; hence, all of the results for the latter model that are in Section 2.2 can be used for the former model. Note, also, from (3.10), that the () ( ) v j j ht h t = , , (, ) jl jl frk have been absorbed into . ( ) v j h t 58 Figure 3.5: Virtual Reservoir Model for N injectors and M producers. Definition 3: Let R j IP denotes the regional impact, then, analogous to j IPR in (2.14), 1,,, 0 1 , 1 1 () ( )| ( , ) ( )| M Rv v j j j z jl jl jl z t l M jl l IP h t dt H z f r k H z T IPR ∞ 1 = = = = = =≈ = = ∑ ∫ ∑ (3.11) Note that the right-most side of the first row of (3.11) is obtained using (3.7), when (3.7) is expressed using z-transform, and , jl IPR , which equals to ,, , (, ) ()| jl jl jl z fr k H z 1 = , is obtained using (2.14). 1 ( ) it Virtual Channel 1: 1 ( ) v h t ∑ ,1 ( ) m it 1 ( ) nt 2 ( ) i t ∑ ,2 ( ) m it 2 ( ) nt ( ) N i t ∑ , ( ) mN it ( ) N n t Virtual Channel N: ( ) v N h t Virtual Channel 2: h 2 ( ) v t 1 () v p t Σ ( ) R p t ( ) R p nt 2 () v p t ( ) R m p t Σ ( ) v N p t 59 3.2.3 Pseudo-virtual reservoir model As seen from (3.11), to estimate the regional impact R j IP one needs to estimate the virtual channel impulse response , which is the sum of M AR models having 2 ( ) v j h n M unknown parameters. In order to simplify the estimation process, we reduce the number of unknown parameters by approximating the virtual channel as a product of a pseudo-virtual channel and a pseudo-scale function ( ) v j H z ( ) pv j H z (, pv pv j j ) frk , i.e., (3.12) () ( , ) ( ) vpvpvp jjjj Hz f r k H z ≈ v where 1 12 () (1 ) pv j pv j pv j z Hz z γ α − − = − (3.13) and, and are the pseudo-distance and pseudo-permeability between the injector j and the regional producer. This approximation is made by equating the area under to the area under pv j r pv j k ( ) v j h t (, ) () pv pv pv jj j frk H z . An example is shown in Fig. 3.6, where the solid curve represents , the dash curve represents ( ) v j h t (, ) () pv pv pv jj j frk H z . Using (, ) () pv pv pv jj j frk H s () R P s j , in (3.10) can be approximated, as (3.14) 11 1 () () () () ( , ) ( ) ( ) NN N Rv v pvpvpv jjj jjj jj j Ps P s H sI s f r k H sI s == = == ≈ ∑∑ ∑ 60 This demonstrates that the virtual reservoir model in Fig. 3.4 can be further reformulated as the pseudo-virtual reservoir model depicted in Fig. 3.7. ( ) h • Figure 3.6: An example of the virtual impulse response () v j h t and the scaled pseudo-impulse response (, ) ( pv pv pv jj j ) frk h t . Figure 3.7: Pseudo- virtual reservoir model for N injectors and M producers. t ( ) v j h t (, ) () pv pv pv jj j fr k h t 1 ( ) it 1 ( ) pv ht 11 (, ) pvpv frk ∑ ,1 ( ) m it 1 ( ) nt 2 ( ) i t ∑ ,2 ( ) m it 2 ( ) nt ( ) N i t ∑ , ( ) mN it ( ) N n t 22 (, ) pvpv frk ( ) pv N ht 1 () v p t 2 ( ) pv ht Σ ( ) R p t ( ) R p nt 2 ( ) v p t ( ) R m p t Σ ( ) v N p t (, ) pvpv NN frk 61 Comparing Figs. 3.7 and 2.1, observe that the pseudo-virtual reservoir model is very similar to the reservoir model for multiple injectors and a single producer, as described in Section 2.2, except that: 1) The scale function ( , ) j j f rk is replaced by the pseudo-scale function (, pv pv j j ) frk ; 2) The channel is replaced by the pseudo-virtual channel ; and, ( ) j ht ( ) pv j h t 3) The production rate ( ) p t is replaced by the regional production rate ( ) R p t . Substituting the approximation of (3.12) into (3.12), it follows that the regional impact can be further approxiamted as 1 ()| ( , ) ()| R v pv pv pv jj z j j j IP H z f r k H z 1z = = ≈≈ (3.15) As in (2.14) and (2.15), (3.15) can be expressed as: 1 2 (, ) (, ) ()| (1 ) (1 ) pv pv pv pv jj j j Rpvpv pv jj j j z pv pv jj fr k IP f r k H z γγ αα = 2 ′ ≈= = −− (3.16) where (, ) pv pv pv pv jjj fr k j γ γ ′ = . Consequently, by replacing with () v j ht (, ) () pv pv pv jj j frk h t , the estimation of R j IP has been tremendously simplified because the number of parameters that need to be estimated have been reduced from 2M to 2. In short, the 62 estimation of R j IP for a region becomes the same as the estimation of j IPR for multiple injectors and a single producer. Using the available sampled measurements of injection and production rates, our goals are: 1) Estimate the pseudo-virtual impulse response parameters pv j γ ′ and pv j α ; and, () 1, , j N = … 2) Compute ˆ R j IP using ˆ pv j γ ′ and ˆ pv j α ( ) 1, , jN = … ; These goals will be achieved by using the EKF estimator. 3.2.4 Estimator using Extended Kalman Filter (EKF estimator) A. State variable model (SVM) The derivation of the state variable model (SVM) for this pseudo-virtual reservoir model is similar to that in Section 2.3.1; however, for easy reference, we still explain the derivation process herein but using the notations for the pseudo-virtual channels. To illustrate how to do this, we consider the simple case of three injectors and multiple producers (it is easy to generalize the SVM from three injectors to N injectors.). Observe from (3.9), (3.12) and (3.13) that for each pseudo-virtual channel: 11 12 2 2 () ( , ) () (1 ) 1 2 ( ) vpvpvpv pv jjjj j pv pv pv jj jj Pz f r k z z Izz γγ ααα −− z − − ′ ≈= −−+ (3.17) which can be further expressed as: 63 () (3.18) 22 1 1 2 ( ) () ( ) pv pv v pv jj j j j zP z zI z αα γ −− ′ −+ ≈ Multiply on both sides of (3.18) and determine the inverse z-transform of the resulting equation, to see that (3.18) can be expressed in the time-domain, as: z 2 (1) 2 () ( ) ( 1) ( vpvv pvv pv jjj jj j ) j pkpk pk αα γ ′ +− + − ≈ik (3.19) Because only measured injection rates, , are available, where , ( ) mj i k , () () ( ) mj j j ik ik nk = + , (3.20) it follows that (3.19) becomes: 2 , (1) 2 () ( ) ( 1) () ( vpvv pvv pv pv jjj jj jmj j ) j pkpk pk ik n αα γ γ ′′ +− + − ≈ − k ) j (3.21) Letting , (3.21) simplifies to , () ( ) pv mj j j j nk nk γ ′ ≡− 2 ,, (1) 2 () ( ) ( 1) () ( vpvv pvv pv jjj jj jmj m pkpk pk ikn αα γ ′ +− + − ≈ +k ) (3.22) (3.22) is a second-order finite-difference equation that is described by two state variables and ( 1 v j pk − ( ) v j p k . In addition, the unknown parameters, pv j γ ′ and pv j α are also treated as state variables so that they can be estimated by an EKF, and their state equation models are 64 (1) () () (1) () () pv j pv j pv pv jj pv pv jj kkn kkn α γ αα γγ ′ ⎫ += +k k ⎪ ⎪ ⎬ ⎪ ′′ += + ⎪ ⎭ (3.23) where and are discrete-time zero-mean white random noises. Eqs. (3.22) and (3.23) together are described by a ( ) pv j n k α ( ) pv j n k γ ′ 4 1 × state vector j x , where ( ), ( ), ( 1), ( ) pv pv v v jj j j j k k pk pk αγ ′ ⎡⎤ ′ =− ⎣⎦ x (3.24) Using standard techniques [19], the following SVM is obtained for each ( ) v j p k ( ): 1,2, j = 3 )k (3.25) 1 1 2 2 4 3 2 14 1 3 4 2, () (1) (1) () (1) (1) (1) () ( (1) () 2() () () () (1) (1) () () (1) 0 j pv j j j j pv j j jj v j j jj j j v j j jmj v j xk k xk xk xk k kxk xk pk xkx k xkx k xk pk xki k pk α γ ⎡⎤ ⎡⎤ + + ⎢⎥ ⎡⎤ ⎢⎥ ⎢⎥ ⎢⎥ ′⎢⎥ + + ⎢⎥ ⎢⎥ += = = + ⎢⎥ ⎢⎥ ⎢⎥ + ⎢⎥ ⎢⎥ ⎢⎥ − ⎢⎥ ⎢⎥ ⎢⎥ + + ⎣⎦ ⎢⎥ ⎣⎦⎢⎥ + ⎣⎦ += x xn () 0 0 1 ( 1) j k ⎧ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎨ ⎪ ⎪ ⎪ ⎪ ⎪ + ⎪ ⎩ x where 65 , () [ (), (), 0, ()] pv pv i jj mj kn k n k n k αγ ′ ′ = x n (3.26) is a 41 zero-mean white noise vector. Letting × [ ] 12 3 (1) ( 1), (1), ( 1) kcol k k k +≡ + + + xx x x , the complete SVM for three injectors is: 11 11 12 12 14 13 2 11 14 11 13 12 ,1 14 21 22 23 24 31 32 33 34 () (1) () (1) () (1) 2() () () () () ( (1) (1) (1) (1) (1) (1) (1) (1) (1) (1) m xk xk xk xk xk xk xkx k xkx k x ki k xk xk xk k xk xk xk xk xk xk + ⎡⎤ ⎢⎥ + ⎢⎥ ⎢⎥ + ⎢⎥ −+ + ⎢⎥ ⎢⎥ + ⎢⎥ + ⎢⎥ += = ⎢⎥ + ⎢⎥ ⎢⎥ + ⎢⎥ + ⎢⎥ ⎢⎥ + ⎢⎥ + ⎢⎥ ⎢⎥ + ⎣⎦ x 21 22 24 2 21 24 21 23 22 ,2 31 32 34 2 31 34 31 33 32 ,3 12 3 ) () () () () 2() () () () () () () () () 2() () () () () () (1) ( 1) (1) ( 1) m m Rv v v m xk xk k xk xkx k xkx k x ki k xk xk xk x kxk x kx k xki k p k pk pk pk ⎡⎤ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ + ⎢⎥ ⎢⎥ −+ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ ⎢⎥ −+ ⎣⎦ += + + + + + x n () (1) 0 0 0 1 0 0 0 1 0 0 0 1 ( 1) ( 1) R p p nk knk ⎧ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎨ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ++ ⎪ ⎪ =+ ⎩ x++ (3.27) where is the measured regional production rate, and (1) R m pk + 12 3 () [ ( ), (), ( )] k col k k k = xx x x nnnn ( 1) R p nk + are additive zero-mean white noises 66 for the state and measurement equations with covariance matrix and variance k Q 1 k r + , respectively, where (3.28) ,1 ,2 ,3 11 2 2 3 3 , , 0, , , , 0, , , , 0, mm nn n n n n n n n diag r r r r r r r r r αγ α γ α γ ⎡⎤ = ⎣⎦ k Q m ) ) Because this SVM is nonlinear, we can’t use a Kalman filter to estimate its twelve states; Instead, we use an EKF to estimate them. After all 12 states are estimated, we extract the estimates of the six virtual reservoir model parameters from them, namely and ˆ (1| 1 pv j kk α + + ˆ (1| 1 pv j kk γ ′ + + ( 1, 2, j 3 = ), after which we compute ( ) as ˆ (1| 1 R j IP k k + +) 3 1, 2, j = 2 ˆ (1| 1) ˆ (1| 1) ˆ (1 ( 1| 1)) pv j R j pv j kk IP k k kk γ α ′ ++ ++ = −+ + (3.29) In this way, we see that ˆ (1| 1 R j IP k k ) + + ( 1, 2, j 3 = ) are functions of time, so that the EKF will provide dynamic update values for ˆ R j IP , which can be very useful if there is a change to the reservoir. B. EKF approach The EKF approach herein is exactly the same as described in Section 2.3.2; hence, we omit it, and refer the reader to Section 2.3.2. 67 3.3 Simulations The pseudo-virtual reservoir method will be tested using synthetic data, reservoir simulation data and real data. The tests using synthetic data, described below, evaluate the feasibility of our approach. The tests using reservoir simulation data, to be performed at a later date and not as part of this dissertation, will confirm whether or not our estimated results match the models used for the reservoir simulation data. The tests using real data, also to be performed at a later date and not as part of this dissertation, will confirm whether or not our results match some expert knowledge or actual field test result. Table 3.1: Summarization of the settings for the test cases (For all nine cases, there are two producers in the region). Case # Name Number of close injectors Number of distant injectors Constant or non-constant IPRs Low or high impact from distant injectors 1 Static reservoir case 1 3 0 Constant NA 2 Static reservoir case 2 3 0 Constant NA 3 Dynamic reservoir case 1 3 0 Non-Constant NA 4 Dynamic reservoir case 2 3 0 Non-Constant NA 5 Static reservoir with low impact distant injector case 1 3 1 Constant Low 6 Static reservoir with low impact distant injector case 2 3 1 Constant Low 7 Dynamic reservoir with low impact distant injector 3 1 Non-Constant Low 8 Static reservoir with high impact distant injector 3 1 Constant High 9 Dynamic reservoir with high impact distant injector 3 1 Non-Constant High 68 Our tests using synthetic data are summarized in Table 3.1, according to the number of close injectors, number of distant injectors, constant or non-constant IPRs and low or high impact from distant injectors, e.g., for Case 5, the region has three close injectors and one low impact distant injector with all constant IPRs. 3.3.1 Procedure for choosing the values of ,, (, ) jl jl frk , and , jl a , jl b Recall from our third heuristic-knowledge for , , (, ) jl jl frk in Section 2.1.B, the larger , , (, ) jl jl frk is, the better is the connectivity from the injector to producer; the shorter is the delay time (i.e., the larger is ), and the larger is the impact from the injectors (i.e., the larger is , jl a , jl IPR ). Our procedure for choosing the values of ,, (, ) jl jl frk , is: , , jl jl a b , 1) Specify the value of , , (, ) jl jl frk in [ ] 0, 1 ; 2) Choose the value of [ ] , 1, 1000 jl a ∈ and [ ] , 0, 1 jl IPR ∈ using following rules: Rule 1: If , , (, ) jl jl frk is large, then and , jl a , jl IPR are large. Rule 2: If , , (, ) jl jl frk is moderate, then and , jl a , jl IPR are moderate. Rule 3: If , , (, ) jl jl frk is small, then and , jl a , jl IPR are small. 3) Determine using , jl b , , j l a T jl e α − = , ,, , jl j l jl b T γ α = and (2.8) as 69 () () () ,, ,, ,, 1 , 1 , 22 , ,, , 2 ,, ()| ( )| 1 1 (, )1 jl jl jl jl jl jl z j l z jl aT aT jl jl j l jl aT aT jl jl Hz Hz b T T ee IPR fr k e e T γ α αα == −− −− == = − − = − T (3.33) where T is a small sample period, which corresponds to high sampling rate so that the generated discrete signal can be considered as a continuous signal. The regional impact, R j IP , can then be computed using (3.11), i.e., , 1 M R jj l l IP IPR = = ∑ (3.34) 3.3.2 Synthetic data generation Synthetic data was generated as follows: (Note: the procedure is explained for the case of three injectors and two producers, and can be extended to any other situation.) Step 1: Generate the noise-free injection rates ( ( ) h j i n 1, 2, 3 j = ) using the injection rate model described in Appendix A. Step 2: To generate the noise-free production rates ( ) h l p n (1 ), the SVM for multiple injectors and a single producer was implemented with zero-noise by specifying ,2 l = , , (, ) ij ij frk , ( ,, , ij ij a b 1,2,3; 1,2. i j = = ) using the procedure described in Section 3.1 and a small 0.01 T = , which corresponds to a high sampling rate (recall that and , , ij a T ij e α − = ,, , ij ij ij bT γ α = ). 70 Step 3: The noise-free injection rates ( ( ) j i k 1,2, j 3 = ) and noise-free production rates () l p k (1 ) were obtained by down-sampling ( ) and ,2 l = ( ) h j i n 1,2, i = 3 ( ) h l p n by 20. This corresponds to a low sampling rate. Step 4: The measured injection and production rates were obtained by adding measurement noises with ( ) j nk () () 2 10 2 () 10log 20 () j j Ei k SNR En k ⎡⎤ ⎢ ⎥ = = ⎢⎥ ⎣⎦ dB to ( ), and with ( ) j ik 1, 2, j = 3 ( ) l p n k () () 2 10 2 () 10log 20 () l l p Epk SNR Enk ⎡⎤ ⎡⎤ ⎢⎥ ⎣ ⎦ = = ⎢⎥ ⎡⎤ ⎢⎥ ⎢⎥ ⎣⎦ ⎣⎦ dB to () l p k (1 ), i.e., ,2 l = , , () () () () () () l mj j j ml l p ik ik nk p kpk n k =+ ⎧ ⎪ ⎨ ⎪ =+ ⎩ (3.35) 3.3.3 Monte-Carlo simulations Simulations were implemented as follows: Step 1: Set , number of realizations 1 i = 100 NR = , and the number of samples in each realization 500 M = . Step 2: For the realization ( th i 1, 2, 3 j = , 1, 2 l = , 1, ,500 k = … ): 1) Generate and , ( ) mj i k , ( ) ml p k , using the four-step procedure described in 71 Section 3.3.2, and compute the measured regional production rate ( ) R m p k using (3.1); 2) Implement the EKF using the just generated and , ( ) mj i k ( ) R m p k , and obtain and ; ˆ ( | ) pv j kk γ ′ ˆ (| ) pv j kk α 3) Compute the estimated regional impact ˆ (|) R j i IPk k ⎡ ⎤ ⎣ ⎦ using () 2 ˆ (|) ˆ (| ) ˆ 1(|) pv j R j i pv j kk IP k k kk γ α ′ ⎡⎤ = ⎣⎦ − . (3.36) Step 3: If , go to Step 4; otherwise, set i NR ≥ 1 ii = + and go to Step 2. Step 4: Compute the mean and standard deviation of ˆ (| ) R j i IPk k ⎡ ⎤ ⎣ ⎦ using 1 1 ˆ () ( | ) NR R jj i i mk IP k k NR = ⎡ ⎤ = ⎣ ⎦ ∑ (3.37) and ( ) 2 1 1 ˆ () ( | ) ( ) NR R jj i i kIPkkm NR σ = ⎡⎤ = ⎣⎦ ∑ j k− , (3.38) respectively. Step 5: Stop. 72 Case 1: Static reservoir case 1 The parameters for Case 1 are summarized in Table 3.2. For this case, 1) Injector 1 impacts producer 2 the most while injector 2 impacts producer 1 the most because 1,2 1,2 (, ) 1 fr k = and 2,1 2,1 (, ) 1 fr k = . 2) Injector 3 impacts producers 1 and 2 the least because 3,1 3,1 (, ) 0.5 fr k = and . 3,2 3,2 (, ) 0.5 fr k = 3) Injector 2 has the largest regional impact, after which comes injector 1. Injector 3 has the smallest regional impact. Table 3.2: The parameters for Case 1 (i refers to injector and j refers to producers). i j , ij a , i j b , i j α , ij γ , i j IPR ,, (, ) ij ij frk R j IP 1 200 120 0.1353 0.1624 0.22 0.7 1 2 400 1200 0.0183 0.2198 0.23 1 0.384 1 400 2400 0.0183 0.4396 0.456 1 2 2 200 80 0.1353 0.1083 0.14 0.8 0.576 1 10 0.1 0.9048 0.0009 0.1 0.5 3 2 10 0.05 0.9048 0.0005 0.05 0.5 0.075 The estimation results for regional impacts ˆ (| ) R j IPk k ( 1, 2, 3 j = ) are depicted in Figs. 3.8 and 3.9. Observe that: 1) The means of ˆ (| ) R j IPk k are very close to their actual regional impact as k increases. 2) The standard deviations of ˆ (| ) R j IPk k are very small, which means that the estimation is very stable. 73 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The mean of estimated regional impacts m 1 (k) 0.384 m 2 (k) 0.576 m 3 (k) 0.075 Figure 3.8: The mean of estimated ˆ (| ) R j IPk k for Case 1. 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The standard deviations of estimated regional impacts σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.9: The standard deviation of estimated ˆ (| ) R j IPk k for Case 1. 74 Case 2: Static reservoir case 2 The parameters for Case 2 are summarized in Table 3.3. For this case: 1) Injector 3, instead of injector 1, impacts producer 2 the most because and 3,2 3,2 (, ) fr k =1 1,2 1,2 (, ) 1 fr k = . 2) Injector 1 has a regional impact that is very close to that of injector 3. Table 3.3: The parameters for case 2 (i refers to injector and j refers to producers). i j , ij a , ij b , ij α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 200 120 0.1353 0.1624 0.22 0.7 1 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 400 2400 0.0183 0.4396 0.456 1 2 2 10 0.05 0.9048 0.0005 0.05 0.5 0.485 1 10 0.1 0.9048 0.0009 0.1 0.5 3 2 400 1200 0.0183 0.2198 0.23 1 0.28 Estimation results for regional impacts ˆ ( | ) R j IPk k ( 1, 2, j 3 = ) are depicted in Figs. 3.10 and 3.11, which are similar to those for Case 1, i.e., 1) The means of ˆ (| ) R j IPk k are very close to their actual regional impact as k increases. 2) The standard deviations of ˆ (| ) R j IPk k are very small, which means that the estimation is very stable. 75 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The mean of estimated regional impacts m 1 (k) 0.27 m 2 (k) 0.485 m 3 (k) 0.28 Figure 3.10: The mean of ˆ ( | ) R j IPk k for Case 2. 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The standard deviations of estimated regional impacts σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.11: The standard deviation of ˆ ( | ) R j IPk k for Case 2. 76 Case 3: Dynamic reservoir case 1 Unlike Cases 1 and 2, for which the IPRs were constant for all measurements, in this case, we assume that something such as a work-over has happened so that the reservoir model is different before and after 167 k = . The parameters for this case are summarized in Tables 3.4 and 3.5. Observe that: 1) Before , injectors 2 and 3 have the largest and smallest impacts on the region, respectively. After 167 k = 167 k = , injector 1 becomes most influential on producer 1 and least influential on producer 2, whereas injector 2 becomes least influential on producer 1. 2) Injector 2 has the largest regional impact before 167 k = , whereas after 167 k = , injector 1 has the largest regional impact. Estimation results for regional impacts ˆ ( | ) R j IPk k ( 1, 2, j 3 = ) are depicted in Figs. 3.12 and 3.13. Observe that: 1) ˆ ( | ) R j IPk k changes at time 167 k = , which suggests that our method can successfully track a system change. 2) The means of ˆ (| ) R j IPk k are very close to actual regional impacts, both before and after . 167 k = 3) The standard deviations of ˆ (| ) R j IPk k are very small, which means that the estimation is very stable. 77 Table 3.4: The parameters for Case 3 when 1, ,167 k = … (i refers to injector and j refers to producers). i j , ij a , ij b , ij α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 200 120 0.1353 0.1624 0.22 0.7 1 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 400 2400 0.0183 0.4396 0.456 1 2 2 400 1200 0.0183 0.2198 0.23 1 0.686 1 10 0.1 0.9048 0.0009 0.1 0.5 3 2 10 0.05 0.9048 0.0005 0.05 0.5 0.075 Table 3.5: The parameters for Case 3 when 168, ,500 k = … (i refers to injector and j refers to producers). i j , ij a , ij b , ij α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 400 2400 0.0183 0.4396 0.456 1 1 2 10 0.05 0.9048 0.0005 0.05 0.5 0.481 1 10 0.1 0.9048 0.0009 0.1 0.5 2 2 400 1200 0.0183 0.2198 0.23 1 0.28 1 200 120 0.1353 0.1624 0.22 0.7 3 2 200 80 0.1353 0.1083 0.14 0.8 0.27 78 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The mean of estimated regional impacts m 1 (k) 0.27, 0.48 m 2 (k) 0.686, 0.28 m 3 (k) 0.075, 0.27 Figure 3.12: The mean of ˆ ( | ) R j IPk k for Case 3. 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The standard deviations of estimated regional impacts σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.13: The standard deviation of ˆ ( | ) R j IPk k for Case 3. 79 Case 4: Dynamic reservoir case 2 As in Case 3, we assume that something such as a work-over has happened so that the reservoir model is different before and after 167 k = . The parameters for this case are summarized in Table 3.6 and 3.7. Observe that: 1) Before , injector 1 impacts producer 1 the most, injector 2 impacts producer 2 the most, and injector 3 impacts producers 1 and 2 the least, respectively. After 167 k = 167 k = , injectors 1 and 2 impact producers 2 and 1 the least, respectively. 2) Before , injector 2 has moderate regional impact, which is much greater than that from injector 3. After 167 k = 167 k = , it is close to that of injector 3. Table 3.6: The parameters for Case 4 when 1, ,167 k = … (i refers to injector and j refers to producers). i j , ij a , ij b , ij α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 400 2400 0.0183 0.4396 0.456 1 1 2 200 80 0.1353 0.1083 0.14 0.8 0.576 1 200 120 0.1353 0.1624 0.22 0.7 2 2 400 1200 0.0183 0.2198 0.23 1 0.384 1 10 0.1 0.9048 0.0009 0.1 0.5 3 2 10 0.05 0.9048 0.0005 0.05 0.5 0.075 Table 3.7: The parameters for Case 4 when 168, ,500 k = … (i refers to injector and j refers to producers). i j , ij a , ij b , ij α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 400 2400 0.0183 0.4396 0.456 1 1 2 10 0.05 0.9048 0.0005 0.05 0.5 0.481 1 10 0.1 0.9048 0.0009 0.1 0.5 2 2 400 1200 0.0183 0.2198 0.23 1 0.28 1 200 120 0.1353 0.1624 0.22 0.7 3 2 200 80 0.1353 0.1083 0.14 0.8 0.27 80 Estimation results for regional impacts ˆ ( | ) R j IPk k ( 1, 2, j 3 = ) are depicted in Figs. 3.14 and 3.15, and are very similar to those for Case 3, i.e., 1) ˆ ( | ) R j IPk k changes at time 167 k = , which suggests that our method can successfully track a system change. 2) The means of ˆ (| ) R j IPk k are very close to actual regional impact, both before and after . 167 k = 3) The standard deviations of ˆ (| ) R j IPk k are very small, which means that the estimation is very stable. 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The mean of estimated regional impacts m 1 (k) 0.576, 0.48 m 2 (k) 0.384, 0.28 m 3 (k) 0.075, 0.27 Figure 3.14: The mean of ˆ ( | ) R j IPk k for Case 4. 81 0 50 100 150 200 250 300 350 400 450 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 k The standard deviations of estimated regional impacts σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.15: The standard deviation of ˆ ( | ) R j IPk k for Case 4. Case 5: Static reservoir with low impact distant injector case 1 Unlike Cases 1 to 4, beginning with this case, we introduce a distant injector into the region. The purpose of this experiment is to see if omitting the distant injector from the EKF makes a difference on the estimated regional impact for close injectors. The corresponding parameters for this case are summarized in Table 3.8. Observe that: 1) The region includes three close injectors 1 to 3, one distant injector 4, and two producers, and the distant injector has very little impact on the producers. 2) Among the three close injectors, injector 2 has the largest regional impact whereas injector 3 has the smallest regional impact. 82 Estimation results for regional impacts ˆ ( | ) R j IPk k ( 1, 2, j 3 = ) using injection rates from only the three close injectors are depicted in Figs. 3.16 and 3.17, and estimation results using injection rates from all four injectors are depicted in Figs. 3.18 and 3.19, respectively. Observe that: 1) The means and standard deviations of estimated regional impacts ˆ (| ) R j IPk k for close injectors, without using the distant injector, are very close to those using all four injectors. 2) Regardless of using or not using the distant injector, the means of ˆ (| ) R j IPk k for the three close injectors are very close to the actual regional impacts as k increases. 3) Regardless of using or not using the distant injector, the standard deviations of ˆ ( | ) R j IPk k for the three close injectors are very small, which means that the estimation is very stable. 4) ˆ ( | ) R j IPk k for the distant injector is very small, and this agrees with the reservoir model. Table 3.8: The parameters for Case 5 (i refers to injector and j refers to producers). i j , ij a , ij b , i j α , ij γ , i j IPR ,, (, ) ij ij frk R j IP 1 200 120 0.1353 0.1624 0.22 0.7 1 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 400 2400 0.0183 0.4396 0.456 1 2 2 400 1200 0.0183 0.2198 0.23 1 0.686 1 10 0.1 0.9048 0.0009 0.1 0.5 3 2 10 0.05 0.9048 0.0005 0.05 0.5 0.065 1 1 0.0005 0.99 6 510 − × 0.05 0.3 4 2 1 0.0005 0.99 6 510 − × 0.05 0.3 0.03 83 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27 m 2 (k) 0.686 m 3 (k) 0.065 Figure 3.16: The mean of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 5. 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.17: The standard deviation of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 5. 84 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27 m 2 (k) 0.686 m 3 (k) 0.065 m 4 (k) 0.03 Figure 3.18: The mean of ˆ (| ) R j IPk k using injection rates from all injectors for Case 5. 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) σ 4 (k) Figure 3.19: The standard deviation of ˆ ( | ) R j IPk k using injection rates from all injectors for Case 5. 85 Case 6: Static reservoir with low impact distant injector case 2 This Case is similar to Case 5 except that the impacts of close injectors for Case 6 are different from those for Case 5. The parameters for this case are summarized in Table 3.9. Estimation results using injection rates from only the three close injectors are depicted in Figs. 3.20 and 3.21 for estimated regional impacts ˆ ( | ) R j IPk k ( 1, 2, j 3 = ), and estimation results using injection rates from all four injectors are depicted in Figs. 3.22 and 3.23 for estimated regional impacts ˆ ( | ) R j IPk k ( 1,2,3,4 j = ), respectively. All results are similar to those for Case 5, which further confirms the results obtained in Case 5. Table 3.9: The parameters for Case 6 (i refers to injector and j refers to producers). i j , ij a , ij b , i j α , ij γ , i j IPR ,, (, ) ij ij frk R j IP 1 200 120 0.1353 0.1624 0.22 0.7 1 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 400 2400 0.0183 0.4396 0.456 1 2 2 10 0.05 0.9048 0.0005 0.05 0.5 0.485 1 10 0.1 0.9048 0.0009 0.1 0.5 3 2 400 1200 0.0183 0.2198 0.23 1 0.28 1 1 0.0005 0.99 6 510 − × 0.05 0.3 4 2 1 0.0005 0.99 6 510 − × 0.05 0.3 0.03 86 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27 m 2 (k) 0.485 m 3 (k) 0.28 Figure 3.20: The mean of the estimated regional impact ˆ (| ) R j IPk k using injection rates from only close injectors for Case 6. 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.21: The standard deviation of the estimated regional impact ˆ ( | ) R j IPk k using injection rates from only close injectors for Case 6. 87 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27 m 2 (k) 0.485 m 3 (k) 0.28 m 4 (k) 0.03 Figure 3.22: The mean of the estimated regional impact ˆ (| ) R j IPk k using injection rates from all injectors for Case 6. 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) σ 4 (k) Figure 3.23: The standard deviation of the estimated regional impact ˆ ( | ) R j IPk k using injection rates from all injectors for Case 6. 88 Case 7: Dynamic reservoir with low impact distant injector Unlike Case 6, in this case, we assume that something such as a work-over has happened so that the reservoir model is different before and after . The parameters for this case are summarized in Tables 3.10 and 3.11. Observe that: 167 k = 1) Before , injector 2 impacts producers 1 and 2 the most, and injector 3 impacts producers 1 and 2 the least. After 167 k = 167 k = , injector 1 impacts producer 1 the most and producer 2 the least, and injector 2 impacts producer 2 the most and producer 1 the least. 2) Before , injector 2 has the largest regional impact, whereas after , injector 1 has the largest regional impact. 167 k = 167 k = Estimation results for regional impacts ˆ ( | ) R j IPk k ( 1, 2, 3 j = ) using injection rates from only the three close injectors are depicted in Figs. 3.24 and 3.25, and estimation results using injection rates from all four injectors are depicted in Figs. 3.26 and 3.27. Observe that: 1) The estimated regional impacts ˆ (| ) R j IPk k for the three close injectors, without using distant injector, are very close to those using all four injectors. 2) Regardless of using or not using the distant injector, the means of ˆ (| ) R j IPk k for the three close injectors are very close to the actual regional impacts as k increases. 3) Regardless of using or not using distant injector, the ˆ ( | ) R j IPk k changes at time , which verify that our method can adaptively track a system change. 167 k = 89 4) Regardless of using or not using distant injector, the standard deviations of ˆ ( | ) R j IPk k for the three close injectors are very small, which means that the estimation is very stable. 5) ˆ ( | ) R j IPk k for the distant injector is very small, and this agrees with the reservoir model. Table 3.10: The parameters for Case 7 when 1, ,167 k = … (i refers to injector and j refers to producers). i j , ij a , i j b , i j α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 200 120 0.1353 0.1624 0.22 0.7 1 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 400 2400 0.0183 0.4396 0.456 1 2 2 400 1200 0.0183 0.2198 0.23 1 0.686 1 10 0.1 0.9048 0.0009 0.1 0.5 3 2 10 0.05 0.9048 0.0005 0.05 0.5 0.075 1 1 0.0005 0.99 6 510 − × 0.05 0.1 4 2 1 0.0005 0.99 6 510 − × 0.05 0.5 0.03 Table 3.11: The parameters for Case 7 when 168, ,500 k = … (i refers to injector and j refers to producers). i j , ij a , ij b , ij α , ij γ , i j IPR ,, (, ) ij ij frk R j IP 1 400 2400 0.0183 0.4396 0.456 1 1 2 10 0.05 0.9048 0.0005 0.05 0.5 0.481 1 10 0.1 0.9048 0.0009 0.1 0.5 2 2 400 1200 0.0183 0.2198 0.23 1 0.28 1 200 120 0.1353 0.1624 0.22 0.7 3 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 1 0.0005 0.99 6 510 − × 0.05 0.1 4 2 1 0.0005 0.99 6 510 − × 0.05 0.5 0.03 90 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27, 0.48 m 2 (k) 0.69, 0.28 m 3 (k) 0.075, 0.27 Figure 3.24: The mean of the estimated regional impact ˆ (| ) R j IPk k using injection rates from only close injectors for Case 7. 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.25: The standard deviation of the estimated regional impact ˆ ( | ) R j IPk k using injection rates from only close injectors for Case 7. 91 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27, 0.48 m 2 (k) 0.69, 0.28 m 3 (k) 0.075, 0.27 m 4 (k) 0.03, 0.03 Figure 3.26: The mean of the estimated regional impact ˆ (| ) R j IPk k using injection rates from all injectors for Case 7. 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) σ 4 (k) Figure 3.27: The standard deviation of the estimated regional impact ˆ ( | ) R j IPk k using injection rates from all injectors for Case 7. 92 Case 8: Static reservoir with high impact distant injector Unlike Cases 5 and 6, for which the distant injector has the smallest regional impact, beginning with this case, the distant injector has the greatest regional impact. The parameters for this case are summarized in Table 3.12. Table 3.12: The parameters for Case 8 (i refers to injector and j refers to producers). i j , ij a , ij b , i j α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 200 120 0.1353 0.1624 0.22 0.7 1 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 10 0.1 0.9048 0.0009 0.1 0.5 2 2 400 1800 0.0183 0.3297 0.34 1 0.39 1 1 0.0005 0.99 6 510 − × 0.05 0.3 3 2 10 0.05 0.9048 0.0005 0.05 0.3 0.03 1 400 2400 0.0183 0.4396 0.456 1 4 2 1 0.0005 0.99 6 510 − × 0.05 0.3 0.471 Estimation results for regional impacts ˆ ( | ) R j IPk k ( 1, 2, j 3 = ) using injection rates from only the three close injectors are depicted in Figs. 3.28 and 3.29, and estimation results using injection rates from all four injectors are depicted in Figs. 3.30 and 3.31. Observe that: 1) The estimated regional impacts ˆ (| ) R j IPk k for the three close injectors, without using the distant injector, are much different from those using the distant injector. Interestingly, even when the distant significant injector is omitted, the estimated ˆ ( | ) R j IPk k for close injectors still preserve their relative importance, i.e., the injector with the largest regional impact has the greatest value of ˆ ( | ) R j IPk k , and the injector with the smallest regional impact ˆ ( | ) R j IPk k has the smallest value of ˆ ( | ) R j IPk k . 93 2) Using the distant injector, the means of ˆ (| ) R j IPk k are very close to actual regional impacts as k increases. 3) Using the distant injector, the standard deviations of ˆ (| ) R j IPk k are very small, which means that the estimation is very stable. 4) Using the distant injector, ˆ ( | ) R j IPk k for the distant injector has the largest value, and this agrees with the reservoir model. 5) The standard deviations of ˆ ( | ) R j IPk k for the three close injectors using the distant injector are 50% smaller than those without using the distant injector. 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27 m 2 (k) 0.39 m 3 (k) 0.03 Figure 3.28: The mean of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 8. 94 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.29: The standard deviation of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 8. 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27 m 2 (k) 0.39 m 3 (k) 0.03 m 4 (k) 0.47 Figure 3.30: The mean of ˆ (| ) R j IPk k using injection rates from all injectors for Case 8. 95 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) σ 4 (k) Figure 3.31: The standard deviation of ˆ (| ) R j IPk k i ZI ∧ using injection rates from all injectors for Case 8. Case 9: Dynamic reservoir with high impact distant injector Unlike Case 8, for which the IPRs are constant, in this case, we assume that something such as a work-over has happened so that the reservoir model is different before and after . The parameters for this case are summarized in Tables 3.13 and 3.14. Observe that, among the three close injectors, before , injectors 2 and 3 have the largest and smallest regional impacts, respectively, and after , the regional impact of injector 2 reduces to the medium. 167 k = 167 k = 167 k = Estimation results for regional impacts ˆ ( | ) R j IPk k ( 1, 2, j 3 = ) using injection rates from only the three close injectors are depicted in Figs. 3.32 and 3.33, and estimation 96 results using injection rates from all four injectors are depicted in Figs. 3.34 and 3.35, respectively. All results are similar to those for Case 8 except that, regardless of using or not using distant injector, the ˆ (| ) R j IPk k changes at time , which further confirms that our method can adaptively track a system change. 167 k = Table 3.13: Parameters for Case 9 when 1, ,167 k = … (i refers to injector and j refers to producers). i j , ij a , ij b , i j α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 200 120 0.1353 0.1624 0.22 0.7 1 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 10 0.1 0.9048 0.0009 0.1 0.5 2 2 400 1800 0.0183 0.3297 0.34 1 0.39 1 1 0.0005 0.99 6 510 − × 0.05 0.3 3 2 10 0.05 0.9048 0.0005 0.05 0.3 0.03 1 400 2400 0.0183 0.4396 0.456 1 4 2 1 0.0005 0.99 6 510 − × 0.05 0.3 0.471 Table 3.14: The parameters for Case 9 when 168, , 500 k = … (i refers to injector and j refers to producers). i j , ij a , ij b , i j α , ij γ , ij IPR ,, (, ) ij ij frk R j IP 1 10 0.1 0.9048 0.0009 0.1 0.5 1 2 400 1800 0.0183 0.3297 0.34 1 0.39 1 200 120 0.1353 0.1624 0.22 0.7 2 2 200 80 0.1353 0.1083 0.14 0.8 0.27 1 1 0.0005 0.99 6 510 − × 0.05 0.3 3 2 10 0.05 0.9048 0.0005 0.05 0.3 0.03 1 400 2400 0.0183 0.4396 0.456 1 4 2 1 0.0005 0.99 6 510 − × 0.05 0.3 0.471 97 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27, 0.39 m 2 (k) 0.39, 0.27 m 3 (k) 0.03, 0.03 Figure 3.32: The mean of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 9. 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) Figure 3.33: The standard deviation of ˆ (| ) R j IPk k using injection rates from only close injectors for Case 9. 98 0 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The mean of estimated regional impact m 1 (k) 0.27, 0.39 m 2 (k) 0.39, 0.27 m 3 (k) 0.03, 0.03 m 4 (k) 0.47, 0.47 Figure 3.34: The mean of ˆ (| ) R j IPk k using injection rates from all injectors for Case 9. 50 100 150 200 250 300 350 400 450 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Time The standard deviation of estimated regional impact σ 1 (k) σ 2 (k) σ 3 (k) σ 4 (k) Figure 3.35: The standard deviation of ˆ ( | ) R j IPk k using injection rates from all injectors for Case 9. 99 3.4 Conclusions This chapter presents a pseudo-virtual reservoir method to forecast the regional impacts between multiple injectors and multiple producers based on measured production and injection rates. Our approach has been tested on synthetic data, and test results suggest that this pseudo-virtual reservoir method is feasible for evaluating regional impacts with multiple injectors and multiple producers. Our approach could be used to determine regional impacts, and may lead to a better understanding of regional reservoir heterogeneity, all of which will assist Chevron engineers to make better decisions under different scales, e.g., a region with a single producer, a collection of producers, a pattern and a collection of patterns. 100 Chapter 4 Background Knowledge for Type-1 and Type-2 Fuzzy Sets In this chapter, we briefly review some basic definitions and theorems for type-1 and type-2 fuzzy sets such that we can easily understand the materials in Chapter 5 [13]-[15], [21], [22], [24], [25], [28], [29]. 4. 1 Type-1 Fuzzy Sets Definition 4.1: A type-1 fuzzy set, denoted A , is characterized by a T1 membership function (MF) ( ) x µ , where x X ∈ and () [0,1] x µ ∈ , i.e., { } = (, ()) | , () [0,1] Ax x x X x µµ ∀∈ ∈ (4.1) Definition 4.2: The cut α − of the T1 fuzzy set, A α , is a crisp set that can be formulated as { } |() Ax x α µ α = ≥ (4.2) Obviously, A α contains all elements whose grades are greater than or equal to a specific value of α . Consequently, A α is a finite collection of elements for the discrete 101 case and an interval of values for the continuous case. An example is depicted in Fig. 4.1, where the type-1 membership function is triangular, and one of its cuts α − contains all the elements on x-axis between the two lower solid dots. x () x µ 1 All Values of x between the α two end-points is the alpha-cut Figure 4.1: A type-1 triangular membership function and one of its alpha-cuts. x α () x µ Figure 4.2: Membership function of ( ) A α . Definition 4.3: Let ( ) A I x α be an indicator function of the crisp set A α , i.e., (4.3) 1, () 0, otherwise A xA Ix α α ∈ ⎧ = ⎨ ⎩ 102 The associated T1 fuzzy set for the cut α − A α , ( ) A α , is defined as { } ()=(, ()) | A AxIx x α αα ∀∈X (4.4) An example is depicted in Fig. 4.2, where the cut α − A α has been raised to level α . Although the T1 MF is the usual way to represent a T1 fuzzy set, cuts α − provide another way as in the following cut α − representation: Theorem 4.1 ( cut α − Representation Theorem for T1 FS): A T1 FS is equal to the union of its associated T1 FSs A ( ) A α for [0, 1] α ∈ , i.e., [0,1] [0,1] () sup ( ) AA A α α α α ∈ ∈ == ∪ (4.5) Unlike the membership function representation, which is very useful for theoretical analyses, the cut α − representation usually not only provides us a good capability for computation, e.g., in computing a fuzzy weighted average, but it also is very useful for theoretical analyses. 4. 2 Type-2 Fuzzy Sets In the shaded area of Fig. 4.3, at a specific value of x , say x′ , there no longer is a single value for the membership function. Instead, the membership function takes on 103 values wherever the dotted line intersects the shaded area. Different amplitude values can be assigned to these points ( ) , () xu x µ ′′ ′ = , which consequently leads to a three- dimensional membership function, called a type-2 membership function. x (, ) A xu µ u x ′ u ′ Figure 4.3: An example of a type-2 membership function. Definition 4.4: At each value of primary variable x, there is an associated secondary variable, u, where u ∈[u L (x),u U (x)] ≡ J x , and u and denote the L (x) u U (x) lower and upper values for u at each x. is a shorthand notation for [ , and is called the primary membership of x, and, . J x u L (x),u U (x)] J x ⊆ [0,1] 104 Definition 4.5: A type-2 fuzzy set, denoted A , is characterized by a type-2 membership function (, ) A x u µ , where , [0,1] x xX u J ∈ ∀∈ ⊆ , i.e., { } =( , ), ( , ) | , [0,1] x A Axu xu x X u J µ ∀∈ ∀ ∈ ⊆ (4.6) in which 0 ( , ) 1 A xu µ≤≤ , and is the primary membership, which is the domain of the secondary membership function. can also be expressed as x J A (4.7) = ( , ) /( , ) [0,1] x x A xXuJ AxuxuJ µ ∈∈ ⊆ ∫∫ where denotes union over all admissible ∫∫ x and . For discrete universes of discourse is replaced by u ∫ ∑ . In Definition 4.5, the first restriction that is consistent with the T1 constraint that 0 u ∈J x = [u L (x),u U (x)] ⊆[0,1] ≤ µ A (x) ≤ 1, i.e., when uncertainties disappear a T2 MF must reduce to a T1 MF, in which case the variable u equals µ A (x) and 0 ≤ µ A (x) ≤ 1. The second restriction that 0 ≤ µ ˜ A (x,u) ≤ 1 is consistent with the fact that the amplitudes of a MF should lie between or be equal to 0 and 1. Definition 4.6: When ( , ) 1 A xu µ = for ∀x ∈X and u ∈J x , then à is called an interval T2 FS (IT2 FS), in which case 105 (4.8) A = 1/(x,u) u∈J x ∫ x∈X ∫ J x = [u L (x),u U (x)] ⊆ [0,1] Note that the third dimension—the secondary grade—of the general T2 FS is no longer needed for an IT2 FS because it conveys no new information, i.e. all secondary grades of an IT2 FS are the same, namely 1. We therefore have two choices: (1) abandon the third dimension entirely and introduce a different notation for an IT2 FS, since the T2 FS notation carries the connotation of a third dimension; or (2) continue to use the notation but in a simpler way than in (4.6). We choose to do the latter so that the reader will be able to easily connect with the existing literature on IT2 fuzzy systems. A A Definition 4.7: As shown in Fig. 4.4, at each value of x, say x x′ = , imagine that the 2D plane whose axes are u and ( , ) A x u µ ′ intersects the 3D type-2 MF A , and, consequently, obtain an intersection, which is called a vertical slice, denoted ( ) A x µ ′ , of , i.e., A (4.9) ( )= (x , ) ( , ) / [0,1] x x AA A uJ xxu xxuuJ µµ µ ′ ′ ∈ ′′ ′ == = ⊆ ∫ The vertical slice () A x µ ′ is also called the secondary membership function, and its domain is the primary membership of x, , where for x J [0, 1] x J ⊆ x X ∀∈ . At each 106 value of x and u, the amplitude of a secondary membership function, µ ˜ A (x,u) , is called the secondary grade. x (, ) A x u µ 107 Figure 4.4: An example of the vertical slice for a particular type-2 membership function. Definition 4.8: The union of all primary memberships, which represents the uncertainty in the primary memberships of a type-2 fuzzy set , is called the Footprint of uncertainty (the shaded region in Fig. 4.3), denoted F A OU( ) A , i.e., (4.10) FOU( ) x xX AJ ∈ = ∪ 1 α u The Vertical Slice at one particular value of x An IT2 FS is completely described by its footprint of uncertainty (FOU), and an IT2 FS can therefore be represented in the following simpler way: 1()1 x xX AFOUA J ∀∈ = = ∪ (4.11) where a secondary grade of 1 has been put for at all elements in the . ( ) FOU A Definition 4.9: The upper membership function (UMF) and lower membership function (LMF) of à are two T1 MFs that bound the FOU. The UMF is associated with the upper bound of FOU and is denoted ( ˜ A ) µ ˜ A (x) , ∀x∈X, and the LMF is associated with the lower bound of and is denoted FOU( ˜ A ) µ ˜ A (x), ∀x∈X, i. e. 1 µ ˜ A (x) ≡ FOU( ˜ A ) ∀x ∈X (4.12) µ ˜ A (x) ≡ FOU( ˜ A ) ∀x ∈X (4.13) Using this notation, we can re-express the in (4.10), as FOU( ˜ A ) () [ ( ), ( )] A A xX FOU A x x µµ ∀∈ = ∪ (4.14) 1 In the T2 FS literature, notations ( ) A x µ and ( ) A x µ are used instead of u and , respectively. Obviously, these notations are interchangeable. We use the former notation, again so that readers can easily connect with other articles in the IT2 FS literature. L (x) u R (x) 108 As an example, consider a Gaussian primary MF having a fixed mean, m, and an uncertain standard deviation that takes on values in [σ 1 ,σ 2 ] , i.e. 2 1 2 ( ) exp{ [( ) / ] } A x xm µσ =− − , where σ ∈[σ 1 ,σ 2 ]. Corresponding to each value of σ we get a different MF curve. The lower and upper MFs of this IT2 FS are 2 1 1 2 ( ) exp{ [( ) / ] } A x xm µσ =− − and 2 1 2 2 ( ) exp{ [( ) / ] } A x xm µ =− − σ , respectively. Many other examples of IT2 FSs and their lower and upper MFs can be found, e.g. in [21]. Definition 4.10: As shown in Fig. 4.5, for discrete universes of discourse X and , an embedded type-2 fuzzy set contains exactly one element from , namely, , each with its associated secondary grade, namely x J 12 ,, , N xx x JJ J … 12 ,, , N uu u … 11 2 2 ( , ), ( , ), , ( , ) N N AA A xuxu xu µ µ µ … , i.e., 1 (, )/ / N eiii A i Axuu µ = i x ⎡ ⎤ = ⎣ ⎦ ∑ (4.15) An embedded type-1 fuzzy set contains exactly one element from , namely, , i.e., e A 12 ,, , N xx x JJ J … 12 ,, , N uu u … (4.16) 1 / N e i i A u = = ∑ i x 109 Note that acts as the domain for . When each u has been discretized into levels, there are a total of embedded T1 FSs, whereas if each u has been discretized into the same number of M levels, then there are a total of e A e A i M i M i i=1 N ∏ i M N embedded T1 FSs. An example of an embedded T1 FS is depicted in Fig. 4.5; it is the wavy curve 2 on the x u − plane. x (, ) A xu µ u An embedded type-2 fuzzy set An embedded type-1 fuzzy set Figure 4.5: An example of embedded type-2 and type-1 fuzzy set. Theorem 4.2 (Vertical Slice Representation Theorem for T2 FS): The first representation of a T2 FS in the T2 FS literature is the vertical slice representation, in 2 We show the embedded T1 FS as a continuous curve because it is easy to draw such a curve; however, the actual embedded T1 FS, as given by (4.16), is obtained from this curve by sampling it at N values of the primary variable x. 110 which a type-2 fuzzy set A is represented as the union of its vertical slices, i.e., for the continuous case, = ( ,) /( ,) = [ ( , ) / ]/ ( ) / x x A xXuJ AA xX u J xX Axuxu x xu u x x x µ µµ ∈∈ ′′ ∈∈ ∈ ′ ′′ == ∫∫ ∫∫ ∫ ′ N i (4.17) and for the discrete case 11 =( , ) / / = ( , ) / / ( ) / x x i A xX u J N ii i AA iuJ i Axuux xuu x x x µ µµ ∈∈ =∈ = ⎡⎤ ⎣⎦ ⎡⎤ = ⎢⎥ ⎣⎦ ∑∑ ∑∑ ∑ . (4.18) The vertical slice representation is very useful for developing practical computation algorithms. Theorem 4.3 (Embedded Set Representation Theorem for T2 FS): The second representation of a T2 FS is the embedded set representation, in which a general type-2 fuzzy set A is represented as the union of all of its embedded type-2 fuzzy sets, i.e., 1 = n j e j AA = ∑ (4.19) 111 This representation is very useful for theoretical analyses to develop the general structure of the solution to a problem. More important, this representation provides us great understanding about the relationship between type-1 fuzzy sets and type-2 fuzzy sets, as shown in [24]. (4.19) is also called a wavy slice representation of ˜ A . From (4.10), (4.15), (4.16) and (4.19) we see that can be obtained as the union of all of the T1 embedded sets, i.e. ( ) FOU A { } 1 ( ),..., ( ) () (), ( ) A n d A A j e j A A x xxX FOU A A x xx µµ µµ = ⎧ ∀∈ ⎪ == ⎨ ⎡⎤ ∀∈ ⎪ ⎣⎦ ⎩ ∑ X (4.20) The top line of (4.20) is for a discrete universe of discourse, X , and contains n elements (functions), and the bottom line is for a continuous universe of discourse and is an interval set of functions, meaning that it contains an uncountable number of functions that completely fills the space between d A ( ) A x µ and ( ) A x µ , for ∀x ∈X . Equation (4.20) is a wavy-slice representation of , because all of its elements are functions, i.e. they are wavy-slices. ( ) FOU A 112 113 Chapter 5 Encoding Words into Interval Type-2 Fuzzy Sets Using an Interval Approach 5.1 Introduction As mentioned in Chapter 1, the decisions about water-flood management involve a range of activities, some of which are performed only once or twice during the life of the water-flood, e.g., water-flood design and facilities/well planning. Other water-flood activities are performed periodically based on analyses of recurring measurement data, e.g., setting water injection flow rates in the injection wells. How to make decisions based on available information, e.g., the IPRs, is a very difficult problem, especially for large-scale complex system. One possible way to deal with this problem is to construct a decision making system using computing with words (CWW), a methodology, proposed by Zadeh [51]-[52], in which the objects of computation are words and propositions drawn from a natural language. CWW is fundamentally different from the traditional expert systems that are tools to realize an intelligent system but are not able to process natural language because it is imprecise, uncertain and partially true. A specific architecture for CWW using IT2 FSs, called a Perceptual Computer (Per- C), proposed in [23], is shown in Fig. 5.1. The Per-C consists of three components: encoder, CWW engine and decoder. The encoder transforms linguistic perceptions, i.e., words, into IT2 FSs that activate a CWW engine. A CWW engine maps its input IT2 FSs into output IT2 FSs, and this can be done in many different ways, including by rules [21], linguistic summarizations [9], [49], linguistic weighted average [46], etc. The decoder maps the IT2 FS outputs of the CWW engine into a specific word [47]. CWW Engine Encoder Decoder IT2FS IT2FS Perceptions (Words) Perceptions (Words) Figure 5.1: The Perceptual Computer (Per-C) for CWW using IT2 FSs. In order to construct a rule based decision support system using the Per-C, three issues need to be addressed: 1) Design a way to extract rules. Usually, there are two ways to obtain the rules, learning from data and mining from experts using IF-THEN questionnaires. An example of the rules mined from experts may be: : IF the - is , and the between producer and injector is , THEN this injector and producer are within the same pattern. l R injector producer relation small fracture highly impossible highly not 114 115 The rule antecedents (the IF parts of a rule) injector-producer relation and fracture involve the forecasted injector-producer relationships using our newly developed Extended Kalman Filtering approach, and the geology information about the fracture; and each rule consequent (the THEN part of a rule), e.g., highly not, is a word to describe the pattern information. Each expert may provide some rules, and then for a group of experts we can obtain some amount of rules to support decision making. 2) Obtain the membership function for the words such as small, highly impossible, i.e., encoding problem. 3) Map the IT2 FS outputs of the CWW engine into a specific word [27], i.e., decoder problem. This chapter is about the encoding problem, i.e., how to transform a word into an IT2 FS. A collection of application-specific words and their footprints of uncertainty (FOUs) (see Def. 4.8) is a codebook for the application. The codebook is also needed for the decoder. How to collect data from a group of subjects, and how to then map that data into the parameters of a T1 MF has been reported on by a number of authors [e.g., Klir and Yuan (1995)] Names for the different T1 methods include: polling, direct rating, reverse rating, interval estimation, and transition interval estimation. Unfortunately, none of these methods transfers the uncertainties about collecting word data from a group of subjects into the MF of a T1 FS, because a T1 FS does not have enough degrees of freedom to do this; hence, they are not elaborated upon in this thesis. 116 Recently [30], two approaches have been described for collecting data about a word from a group of subjects and then mapping that data into an FOU for that word, the person membership function (MF) approach and the interval end-points approach. Both approaches have been referred to as type-2 fuzzistics [30]. In the person MF approach a subject provides their FOU for a word on a prescribed scale (e.g., 0-10), and this is done for a group of subjects. Each person FOU captures the intra-level of uncertainty about a word, i.e. the uncertainty that each subject has about the word. All of the person-FOUs are aggregated, which captures the inter-level of uncertainty about the word across the group of subjects. Finally, an IT2 FS model is fit to the aggregated data. Note that when the aggregation operation is the union then this approach is based on the T2 FS Representation Theorem (See Theorem 4.3) for an IT2 FS, which states that the FOU of an IT2 FS equals the union of all of its embedded T1 FSs. Each subject’s person FOU can also be interpreted as a union of its embedded T1 FSs. Strong points of this approach are: (1) The union of the person FOUs (the data) establishes the shape of the FOU directly; (2) No statistics about the data are used, i.e. all of the data (the person FOUs) are used so that no information is lost; (3) An a priori uncertainty measure of an FOU is not required in order to map the person FOUs into the IT2 FS that is fit to the aggregated data; and, (4) If all uncertainty disappears then the IT2 FS word model reduces to a T1 FS word model 3 . The weak point of this approach is it requires subjects to be knowledgeable about fuzzy sets. Unfortunately, this weakness may be so large that (in the opinion of the 3 In this case, all subjects provide the same person T1 MF (not an FOU). 117 authors) it may obliterate the advantages of the approach; hence, the person FOU approach is very limited in applicability. In the interval end-points approach each subject provides the end-points of an interval associated with a word on a prescribed scale. Subjects are asked: On a scale of 0-10, what are the end-points of an interval that you associate with the word W? The mean and standard deviation are then computed for the two end-points using the data collected from all of the subjects. Doing this captures the inter-level uncertainties about the word. These end-point statistics are then mapped into an IT2 FS model for the word by bounding the end-points of the centroid of a pre-specified FOU. Mendel and Wu [26] have shown that the centroid of an IT2 FS is a measure of the uncertainty of such a FS, and that, although the centroid cannot be computed in closed form [10], centroid bounds 4 can be expressed explicitly in terms of the geometry (parameters) of an FOU. They provide formulas for these bounds for many different FOUs [26]. Design equations are then used to map the end-point statistics into FOU parameters. Explicit mappings have been provided only for symmetrical FOUs [27]. The strong point of this approach is that collecting interval end-point data is not limited to people who are knowledgeable about fuzzy sets; hence, it is presently a preferable way to collect data from people. Weak points of this approach are: (1) Closed-form mappings are only available for symmetrical FOUs that are associated with data intervals whose two end-point standard deviations are approximately equal, whereas actual interval end-point data show that most words do not have equal end- 4 The centroid is an interval-valued set, and centroid bounds are lower and upper bound for both the left- and right- end points of that set. 118 point standard deviations [20], [21], [30]; (2) The shape of an FOU must be chosen ahead of time, independent of the end-point statistics; (3) The centroid has to be chosen as the measure of uncertainty of an FOU, and because no closed-form formulas are available for the centroid, uncertainty-bound surrogates must be used in their place; (4) Only data end-point mean and standard deviations are used to model word data- uncertainties; and, (5) If all uncertainty disappears (i.e., all subjects provide the same intervals) then the IT2 FS word model does not reduce to a T1 FS word model 5 . Regardless of all of the weaknesses of this approach, its strong point is so large that it has masked the weak points, and has been the recommended approach. In this chapter, a new and simple approach, called the Interval Approach to type-2 fuzzistics is presented, one that captures the strong points (enumerated in Section 5.2) of both the person-MF and interval end-points approaches. The rest of this chapter is organized as follows: Section 5.2 explains the Interval Approach, and provides some observations about the IA; Section 5.3 contains an extensive 32 words codebook example; and, Section 5.4 draws conclusions. Detailed proofs are provided in the appendixes. 5. 2 Interval approach To begin, the Interval Approach (IA) to T2 fuzzistics is overviewed. The IA captures the strong points of both the person-MF and interval end-points approaches, i.e. it: (1) Collects interval end-point data from a group of subjects; (2) Does not require the 5 This last point seems to have been missed when Mendel and Wu [27] set up their two design equations. Data Interval 1 (Subject 1) Data Interval n (Subject n) Data Pre-processing (Four Stages) Data Interval 1 Data Interval m Compute Data Statistics 1 Compute Data Statistics m . . . . . . . . . Assigned Probability Distribution . . . S m S 1 . . . Data Collection Compute Interval Statistics m l and m r m l , m r Figure 5.2: Data Part of the IA approach. Note that the data statistics, , and interval statistics, and , feed into the FS Part of IA, in Fig. 5.5. The extra heavy lines and blocks denote the flow of processing once the data are collected. S 1 ,...,S m m l m r N x 1 Left shoulder FOU Interior FOUs Right shoulder FOU Figure 5.3: Left-shoulder, right-shoulder and interior FOUs, all of whose LMFs and UMFs are piecewise linear. 119 120 subjects to be knowledgeable about fuzzy sets; (3) Has a straightforward mapping from data to an FOU; (4) Does not require an a priori assumption about whether or not an FOU is symmetric or non-symmetric; and, (5) Leads to an IT2 FS word model that reduces to a T1 FS word model automatically if all subjects provide the same intervals. The basic idea of the IA is to map each subject’s data interval into a pre-specified T1 person MF, and to interpret the latter as an embedded T1 FS of an IT2 FS (this is motivated by the Representation Theorem for an IT2 FS. The IA consists of two parts, the Data Part (Fig. 5.2) and the Fuzzy Set (FS) Part (Fig. 5.5). In the Data Part, data that have been collected from a group of subjects are pre-processed, after which data statistics are computed for the surviving data intervals. In the FS Part, FS uncertainty measures are established for a pre-specified T1 MF [always beginning with the assumption that the FOU is an interior FOU (Fig. 5.3), and, if need be, later switching to a shoulder FOU (Fig. 5.3)]. Then the parameters of the T1 MF are determined using the data statistics, and the derived T1 MFs are aggregated using union leading to an FOU for a word, and finally to a mathematical model for the FOU. 5.2.1 Interval approach: data part Once data intervals have been collected from a group of subjects the Data Part of the IA consists of two major steps: (1) Preprocessing the data intervals that have been collected from a group of subjects, and (2) Computing statistics for the data intervals that have survived the pre-processing step. The details of these steps are described in this section. A. Data Preprocessing Preprocessing (Fig. 5.2) the n interval end-point data [ ( consists of four stages: (1) bad data processing, (2) outlier processing, (3) tolerance-limit processing, and (4) reasonable-interval processing. As a result of data preprocessing, some of the n interval data are discarded and the remaining m intervals are re-numbered, . In the rest of this section details are provided for each of these four stages. a (i) ,b (i) ] i = 1,...,n) 1,2,...,m A.1 Stage 1–Bad data processing: Such processing removes nonsensical results (some subjects do not take a survey seriously and provide useless results). If interval end-points satisfy , (5.1) 0 ≤ a (i) ≤ 10 0 ≤ b (i) ≤ 10 b (i) ≥ a (i) ⎫ ⎬ ⎪ ⎭ ⎪ i = 1,...,n then an interval is accepted; otherwise, it is rejected. These conditions are obvious and do not need further explanations. After bad data processing there will be ′ n ≤ n remaining data intervals. A.2 Stage 2–Outlier processing: Such processing uses a Box and Whisker test [43] to eliminate outliers. Recall that outliers are points that are unusually large or small. A Box and Whisker test is usually stated in terms of first and third quartiles and an inter-quartile range. Recall that the first and third quartiles, Q and , (0.25) Q(0.75) 121 contain 25% and 75% of the data, respectively. The inter-quartile range, IQR, is the difference between the third and first quartiles; hence, IQR contains 50% of the data between the first and third quartiles. Any point that is more than 1.5 above the third quartile or more than 1.5 below the first quartile is considered an outlier [43]. Consequently, if the subject interval end-points satisfy IQR IQR (5.2) a (i) ∈[Q a (0.25) − 1.5IQR a ,Q a (0.75) + 1.5IQR a ] b (i) ∈[Q b (0.25) − 1.5IQR b ,Q b (0.75) + 1.5IQR b ] ⎫ ⎬ ⎭ i = 1,..., ′ n a data interval is accepted; otherwise, it is rejected. In these equations, Q (Q ) and ( ) are the quartile and inter-quartile ranges for the left (right) end-points. a b IQR a IQR b After outlier processing, there will be ′ m ≤ ′ n remaining data intervals for which the following data statistics are then computed: m , (sample mean and standard deviation of the left end-points) and , (sample mean and standard deviation of the right end-points). l s l ′ m m r s r ′ m A.3 Stage 3–Tolerance limit processing: If a data interval [ satisfies [43] a (i) ,b (i) ] (5.3) a (i) ∈[m l − ks l ,m l + ks l ] b (i) ∈[m r − ks r ,m r + ks r ] ⎫ ⎬ ⎭ i = 1,..., ′ m it is accepted, otherwise it is rejected. In (5.3) k is determined as follows. 122 Recall that for a normal distribution [43] of measurements with unknown mean and standard deviation, tolerance limits are given by m l ± ks l (or m r ± ks r ), where tolerance factor k is determined so that one can assert with 100(1 − γ )% confidence that the given limits contain at least the proportion 1 − α of the measurements. Table 5.1 (adapted from Table A.7 in 6 [43]) gives k for eight values of ′ m , two values of 1 − γ and two values of 1 − α . Knowing ′ m and choosing values for 1 − γ and 1 − α , one can obtain k. If, e.g., k = 2.549 (for which ′ m = 30) then one can assert with 95% confidence that the given limits contain at least 95% of the subject data intervals. Table 5.1: Tolerance factor k for a number of collected data ( ), a proportion of the data ( ′ m 1− α ), and a confidence level 1− γ [43]. 1− γ = 0.95 1− γ = 0.99 1− α 1− α ′ m 0.90 0.95 0.90 0.95 10 2.839 3.379 3.582 4.265 15 2.480 2.954 2.945 3.507 20 2.310 2.752 2.659 3.168 30 2.140 2.549 2.358 2.841 50 1.996 2.379 2.162 2.576 100 1.874 2.233 1.977 2.355 1000 1.709 2.036 1.736 2.718 ∞ 1.645 1.960 1.645 1.960 Assumption: Data interval end-points are approximately normal, so that the tolerance limits that are given in Table 5.1 can be used. Note that may be different for each word, because it is a result from pre- processing stages 1 and 2, and those stages are applied independently to each word. ′ m 6 Their table is in turn adapted from [2], and contains entries for 47 values of n, beginning with n = 2 . 123 After tolerance limit processing, there will be ′′ m ≤ ′ m remaining data intervals ( 1 ), and the following data statistics are then recomputed: m , (sample mean and standard deviation of the ≤ ′′ m ≤ n l s l ′′ m left end-points) and m , (sample mean and standard deviation of the r s r ′′ m right end-points). A.4 Stage 4–Reasonable-interval processing: In one of our first attempts at the IA [16], only the first three stages of data pre-processing were used. FOUs were obtained that did not look so good (this is subjective, but is demonstrated in Section VI), and many were filled in or almost filled in, i.e. () 0 LMF A ≈ . Because the centroid of a filled-in FOU is completely independent of [32], such an FOU is not considered to be a good one. As a result, something else had to be done. ( ) UMF A It dawned on us that, in addition to focusing on words mean different things to different people (which was our rationale for using IT2 FS models for words), one also needs to focus on words mean similar things to different people. In fact, if there is understanding about a word across a group of subjects, it is the latter that causes it. This led us to require only overlapping intervals be kept. Such intervals are called reasonable, i.e.: Definition 5.1: A data interval is said to be reasonable if it overlaps with another data interval in the sense of Fig. 5.4. 124 a (1) a (2) b (1) b (2) Figure 5.4: An example of two overlapping data intervals for the same word. In the last step of data preprocessing only reasonable data intervals are kept. Appendix B provides a derivation of the following: Reasonable-Interval Test: IF (5.4) a (i) < ξ * b (i) > ξ * ⎫ ⎬ ⎭ ∀i = 1,..., ′′ m where ξ * is one of the values ξ* = (m r σ l 2 − m l σ r 2 ) ± σ l σ r (m l − m r ) 2 + 2(σ r 2 − σ l 2 )ln(σ l / σ r ) ⎡ ⎣ ⎤ ⎦ 1/2 σ l 2 − σ r 2 (5.5) such that m l ≤ ξ* ≤ m r , (5.6) THEN the data interval is kept; OTHERWISE, it is deleted. As a result of reasonable-interval processing, some of the ′′ m data intervals may be discarded and there will finally be m remaining data intervals (1 ) that are re- numbered, 1 . ≤ m ≤ n ,2,...,m 125 B. Computing statistics Two kinds of statistics are computed, one for all of the remaining m data intervals, and one for individual data intervals. Both kinds are used in the Fuzzy Part of the IA. B.1 Compute interval statistics: The sample means of the left and right end-points, and , of the remaining m data intervals are computed. They are used as described in Section 5.2.2.E. m l m r B.2 Compute data statistics for each interval: A probability distribution is assigned to each of the m surviving data intervals after which statistics are computed for each interval using the assumed probability model and the interval end-points. These statistics are used as described in Section 5.2.2.D. Although many choices are possible for an assumed probability distribution for a subject’s data interval, unless a subject provides more information about that interval (e.g., a greater belief in the center of the interval) then a uniform distribution is most sensible, and is the one chosen herein 7 . According to Dubois et al. [3], “ … a uniform probability distribution on a bounded interval … is the most natural probabilistic 126 7 Dubois et al. [3] explain how to map a collection of confidence intervals into a symmetrical triangle T1 MF, where the confidence intervals are associated with data that are collected from a group of subjects about a single point. More specifically, in their problem n measurements, , are collected, after which the sample mean, y 1 , y 2 ,..., y n m y , is computed, as m y = y i / n i=1 n ∑ . The α confidence intervals of m y , denoted [CI(α ),CI(α)], are then computed for a fixed value of α . When each confidence interval is assumed uniformly distributed, their method maps the confidence intervals into a symmetric triangular fuzzy number. Note, however, their problem is different from ours, because in our problem we begin with a collection of n intervals rather than with collection of n numbers; so, their results have not been used by us. representation of incomplete knowledge when only the support is known. It is non- committal in the sense of maximal entropy … and it applies Laplace’s indifference principle stating that what is equipossible is equiprobable.” In order to keep things as simple as possible, only two statistics are used for a uniform distribution, its mean and standard deviation. Recall, that if a random variable Y is uniformly distributed in [ [43], then a,b] m Y = a + b 2 (5.7) σ Y = b − a 12 (5.8) In the second stage of the Data Part, data statistics S are computed for each interval, where 1 ,...,S m S , (5.9) i = (m Y (i) ,σ Y (i) ) i = 1,...,m and these data statistics are then used in the FS part of the IA where they are mapped into the parameters of a T1 MF, as explained in Section 5.2.2.E. This completes the Data Part of the IA. 127 5.2.2 Interval approach: fuzzy set (FS) part The FS Part of the IA (Fig. 5.5) consists of nine steps, each of which is described in this section. A. Choose a T1 FS Model In the present IA, because the mapping from an interval of data to a T1 MF only uses the mean and variance of the (just) assumed uniform probability distribution, only T1 MFs with two degrees of freedom can be used. In this chapter only a symmetrical triangle interior T1 MF, or a left-shoulder T1 MF, or a right-shoulder T1 MF are used. B. Establish FS Uncertainty Measures Although many choices are possible for uncertainty measures of a T1 FS 8 [11] our approach is to focus on simplicity, and to therefore use the mean and standard deviation of a T1 FS. Definition 5.2: The mean and standard deviation of a T1 FS A are: m A = xµ A (x)dx a MF b MF ∫ µ A (x)dx a MF b MF ∫ (5.10) 8 Klir [11] states that there are in principle three uncertainty measures, namely generalized Hartley measure, generalized Shannon measure, and an aggregate measure; however, it has been found by us that these measures are too complicated to use in the IA. 128 σ A = (x − m A ) 2 µ A (x)dx a MF b MF ∫ µ A (x)dx a MF b MF ∫ ⎡ ⎣ ⎢ ⎢ ⎢ ⎤ ⎦ ⎥ ⎥ ⎥ 1/2 (5.11) where and b are the parameters of the MFs that are depicted in the figures of Table 5.2. a MF MF Establish FS Uncertainty Measures Choose TI FS Model Compute Uncertainty Measures For T1 FS Models Compute Embedded T1 FS-1 Compute Embedded T1 FS-m Compute IT2 FS Using Union Compute Mathematical Model . . . . . . . . . IT2 FS Word Model Compute General Formulas for Parameters of T1 FS Models Establish Nature of FOU S m S 1 Delete Inadmissible T1 FSs m l m r Figure 5.5: FS Part of the IA approach. The extra heavy lines and blocks denote the flow of processing once the data are collected. 129 Obviously, if µ A (x) µ A (x)dx a MF b MF ∫ is the probability of x, where , then (5.10) and (5.11) are the same as the mean and standard deviation used in probability. x ∈[a MF ,b MF ] Usually, and b denote the left-end and right-end of the support of a T1 MF; however, shoulder T1 MFs pose a problem because, for a left shoulder T1 MF there is no uncertainty for x , whereas for a right shoulder T1 MF there is no uncertainty for x ; hence, for shoulder MFs a and do not cover the entire span of the MF, and are as shown in the second and third row figures of Table 5.2. a MF MF ∈[0,a MF ] ∈[b MF , M ] MF b MF Table 5.2: Mean and standard deviation for interior and shoulder T1 MFs [16]. Name MF Mean ( ) m MF Standard deviation ( σ MF ) Symmetric triangle (interior MF) a MF b MF x 1 µ(x) m MF = (a MF + b MF )/2 σ MF = (b MF − a MF )/2 6 Left-shoulder a MF b MF x 1 µ(x) m MF = (2a MF + b MF )/ 3 σ MF = 1 6 (a MF + b MF ) 2 + 2a MF 2 ⎡ ⎣ ⎤ ⎦ − m MF 2 ⎡ ⎣ ⎤ ⎦ 1/2 Right-shoulder a MF b MF x 1 µ(x) M m MF = (2a MF + b MF )/ 3 σ MF = 1 6 ( ′ a MF + ′ b MF ) 2 + 2 ′ a MF 2 ⎡ ⎣ ⎤ ⎦ − ′ m MF 2 ⎡ ⎣ ⎤ ⎦ 1/2 ′ a MF = M − b MF ′ b MF = M − a MF ′ m MF == M − m MF 130 C. Compute Uncertainty Measures for T1 FS Models The mean and standard deviations for symmetric triangle (interior), left-shoulder and right-shoulder T1 MFs are easy to compute, and they are also summarized in Table II. Observe that, by using the primed parameters for the right-shoulder T1 MF, the equation for its σ MF looks just like the comparable formula for the left-shoulder T1 MF. D. Compute General Formulas for Parameters of T1 FS Models The parameters of a T1 FS (triangle, left- or right-shoulder) are computed by equating the mean and standard deviation of a T1 FS to the mean and standard deviation, respectively, of a data interval, i.e., m and , where m and are in Table 5.2, and m and are in Table 5.3. This is done for each of the m remaining data intervals. The resulting T1 MF parameters, a and , are summarized in Table 5.3. MF (i) = m Y (i) σ MF (i) = σ Y (i) MF (i) σ MF (i) Y (i) σ Y (i) MF (i) b MF (i) Table 5.3: Transformations of the uniformly distributed data interval [ into the parameters and of a T1 FS [16]. a (i) ,b (i) ] a MF (i) b MF (i) MF Transformations Symmetric triangle (interior MF) a MF (i) = 1 2 [(a (i) + b (i) ) − 2(b (i) − a (i) )] b MF (i) = 1 2 [(a (i) + b (i) ) + 2(b (i) − a (i) )] Left-shoulder a MF (i) = (a (i) + b (i) ) 2 − (b (i) − a (i) ) 6 b MF (i) = (a (i) + b (i) ) 2 + 6(b (i) − a (i) ) 3 Right-shoulder a MF (i) = M − ( ′ a (i) + ′ b (i) ) 2 − 6( ′ b (i) − ′ a (i) ) 3 b MF (i) = M − ( ′ a (i) + ′ b (i) ) 2 + ( ′ b (i) − ′ a (i) ) 6 ′ a (i) = M − b (i) ′ b (i ) = M − a (i) 131 E. Establish Nature of FOU Given a set of m data intervals, they must be mapped into an interior FOU, left- shoulder FOU or a right-shoulder FOU. This is a classification problem. In this sub- section, a rationale for deciding which of these FOUs is chosen, and an FOU classification procedure is explained. E.1 FOU Rationale: To begin, it is always assumed that the m data intervals can be mapped into an interior FOU, and if this cannot be done that the data can be mapped into a left-shoulder FOU, and if this cannot be done that the data can be mapped into a right-shoulder FOU. This rationale is the basis for the classification procedure that is given next. E.2 FOU classification procedure: To begin, the following Admissibility Requirement for an interior FOU is defined. Definition 5.3: For the scale [0, 10], an interior FOU is said to be admissible if and only if (5.12) a MF (i) ≥ 0 b MF (i) ≤ 10 ⎫ ⎬ ⎭ ∀i = 1,..., m By using the formulas for and b that are given in the first row of Table 5.3, it is straightforward to show that (12) is equivalent to: a MF (i) MF (i) 132 , (5.13) 1.207a (i) − 0.207b (i) ≥ 0 1.207b (i) − 0.207a (i) ≤ 10 ⎫ ⎬ ⎭ ∀i = 1,...,m or, equivalently, (5.14) b (i) ≤ 5.831a (i) b (i) ≤ 0.171a (i) + 8.29 ⎫ ⎬ ⎭ ∀i = 1,...,m Additionally, there is the obvious constraint that: b (5.15) (i) ≥ a (i) ∀i = 1,...,m 0 1 2 3 4 5 6 7 8 9 10 8 7 6 5 4 3 2 1 0 Admissible Region for An Interior FOU 9 10 b (i) b (i) ≤ 0.171a (i) + 8.29 b (i) ≥ a (i) a (i) b (i) ≤ 5.831a (i) Figure 5.6: Admissible region for an interior FOU that is based on (5.14) and (5.15). 133 Fig. 5.6 depicts the three inequalities in (5.14) and (5.15) and shows the admissible region for an interior FOU. Unfortunately, requiring (5.14) and (5.15) to be satisfied for all m data intervals is too stringent, as explained next. 0 1 2 3 4 5 6 7 8 9 10 8 7 6 5 4 3 2 1 0 Admissible Region for An Interior FOU 9 10 m b m a m b ≥ m a m b ≤ 0.171m a + 8.29 m b ≤ 5.831m a Figure 5.7: Admissible region for an interior FOU that is based on (5.16) and (5.17). Consider two situations, where in the first situation only one of the m data pairs (barely) falls outside of the admissible region, whereas in the second situation more than half of the data pairs fall outside of that region. Using (5.14) and (5.15) an interior FOU would be rejected for both situations, which does not seem so reasonable; hence, requiring all { to fall in the admissible region seems too stringent. a (i) ,b (i) } i=1 m 134 To that end, instead of using (5.14) and (5.15), their expected values are used, i.e. (5.16) m b ≤ 5.831m a m b ≤ 0.171m a + 8.29 ⎫ ⎬ ⎭ (5.17) m b ≥ m a The figure that is analogous to Fig. 5.6 is now depicted in Fig. 5.7. It looks just like the one in Fig. 5.6, except that the lines are for (5.16) and (5.17) instead of for (5.14) and (5.15). 0 1 2 3 4 5 6 7 8 9 10 8 7 6 5 4 3 2 1 0 Interior FOU 9 10 m r m l m r = m l Left-Shoulder FOU Right-Shoulder FOU Unreasonable Region m r = 5.831m l m r = 0.171m l + 8.29 Figure 5.8: FOU decision regions. 135 In practice, the population-means, m and , are not available. Instead, they are approximated using the interval end-point sample means, m and , which are available from the Data Part of the IA, i.e. a m b l m r (5.18) m a ≈ m l m b ≈ m r ⎫ ⎬ ⎭ Based on these discussions, our FOU Classification Procedure is: Using m and , l m r (5.19) IF m r ≤ 5.831m l , m r ≤ 0.171m l + 8.29, and m r ≥ m l , THEN FOU is an INTERIOR FOU OTHERWISE, IF m r > 5.831m l , and m r < 0.171m l + 8.29, THEN FOU is a LEFT SHOULDER FOU OTHERWISE, IF m r > 0.171m l + 8.29, and m r < 5.831m l RIGHT SHOULDER FOU OTHERWISE, IF m r > 0.171m l + 8.29, and m r > 5.831m l NO FOU ⎫ ⎬ ⎪ ⎪ ⎭ ⎪ ⎪ The FOU decision regions are depicted in Fig. 5.8. Observe that there is a small region for which no FOU is assigned. It is called the Unreasonable Region, because to assign a shoulder FOU for values in it leads to FOUs that extend unreasonably far to the left (for a right-shoulder FOU) or right (for a left-shoulder FOU). No interval data that we have collected have led to that fall in the Unreasonable Region. (m l , m r ) F. Compute Embedded T1 FSs Once a decision has been made as to the kind of FOU for a specific word, each of the word’s remaining m data intervals are mapped into their respective T1 FSs using the 136 equations that are given in Table III, i.e. ( (5.20) a (i) ,b (i) ) → (a MF (i) ,b MF (i) ), i = 1,...,m These T1 FSs, denoted A , are called embedded T1 FSs, because they will be used to obtain the FOU of the word, as described below in Section H. (i) G. Delete Inadmissible T1 FSs It is possible that some of the m embedded T1 FSs are inadmissible, i.e. they violate (5.12). Those T1 FSs are deleted, so that there will be m remaining embedded T1 FSs, where . * m* ≤ m H. Compute an IT2 FS Using the Union Using the Representation Theorem for an IT2 FS, a word’s IT2 FS is computed as: A (5.21) A = A (i) i=1 m* U where is the just-computed i A (i) th embedded T1 FS. I. Compute Mathematical Model for ( ) FOU A In order to compute a mathematical model for , both and ( ) FOU A ( ) UMF A ( ) LMF A must be approximated. There are many ways in which this can be done. Our 137 approach is very simple and guarantees that all m embedded T1 FSs are contained within . Regardless of the type of FOU, the following four numbers must first be computed: * ( ) FOU A a MF ≡ min i=1,...,m* {a MF (i) } a MF ≡ max i=1,...,m* {a MF (i) } ⎫ ⎬ ⎪ ⎭ ⎪ (5.22) b MF ≡ min i=1,...,m* {b MF (i) } b MF ≡ max i=1,...,m* {b MF (i) } ⎫ ⎬ ⎪ ⎭ ⎪ (5.23) I.1 Mathematical model for an interior FOU: Fig. 5.9 depicts this situation. The steps to approximate are: ( ) UMF A a. Compute C MF (i) = a MF (i) + b MF (i) 2 (5.24) b. Compute C MF = min{C MF (i) } (5.25) C MF = max{C MF (i) } (5.26) c. Connect the following points with straight lines: (a MF ,0) , (C MF ,1), (C MF ,1) and (b MF ,0). The result is a trapezoidal UMF. 138 x 1 µ(x) a MF C MF a MF p b MF C MF b MF M µ p Figure 5.9: An example of the union of T1 triangle MFs. The heavy lines are the lower and upper MFs for the interior FOU. x 1 µ(x) a MF a MF b MF b MF Figure 5.10: An example of the union of T1 left-shoulder MFs. The heavy lines are the lower and upper MFs for the left-shoulder FOU. x 1 µ(x) a MF a MF b MF b MF M Figure 5.11: An example of the union of T1 right-shoulder MFs. The heavy lines are the lower and upper MFs for the right-shoulder FOU. 139 The steps to approximate ( ) LMF A are: a. Compute the intersection point (p,µ p ) of the right leg and the left leg of the left and right most-extreme triangles (see Fig. 5.9), using: p = b MF (C MF − a MF ) + a MF (b MF − C MF ) (C MF − a MF ) + (b MF − C MF ) (5.27) µ p = b MF − p b MF − C MF (5.28) b. Connect the following points with straight lines: (a MF ,0) , (a MF ,0), (p, µ p ), (b MF ,0) and (b MF ,0). The result is a triangle LMF. I.2 Mathematical model for a left-shoulder FOU: Fig. 5.10 depicts this situation. To approximate , connect the following points with straight lines: ( , ( ( ) UMF A 0,1) a MF ,1) and (b MF ,0). The result is a left-shoulder UMF. To approximate ( ) LMF A , connect the following points with straight lines: , ( (0,1) a MF ,1), (b MF ,0) and (b MF ,0) . The result is a left-shoulder LMF. I.3 Mathematical model for a right-shoulder FOU: Fig. 5.11 depicts this situation. To approximate , connect the following points with straight lines: ( ( ) UMF A a MF ,0) , 140 (b MF ,1) and (M,1). The result is a right-shoulder UMF. To approximate ( ) LMF A , connect the following points with straight lines: (a MF ,0), (a MF ,0), (b MF ,1) and (M,1) . The result is a right-shoulder LMF. 5.2.3 Observations A. Canonical FOUs for Words Figs. 5.9–5.11 are the only FOUs that can be obtained for a word using the IA, and so these FOUs are referred to herein as canonical FOUs for a word. A word that is modeled by an interior FOU has an UMF that is a trapezoid and a LMF that is a triangle, but in general neither the trapezoid nor the triangle is symmetrical. A word that is modeled as a left- or right-shoulder FOU has trapezoidal upper and lower MFs; however, the legs of the respective two trapezoids are not necessarily parallel. That there are only three canonical FOUs for a word is very different than in function approximation applications of IT2 FSs (e.g., as in fuzzy logic control, or forecasting of time-series) where one is free to choose the shapes of the FOUs ahead of time, and many different choices are possible. B. No Completely Filled-in FOUs In [32], it is explained that when () 0 LMF A = then ( ) FOU A is completely filled-in. This is not considered to be a good FOU, because the centroid of such an FOU equals the span of ( ) LMF A , and is therefore completely independent of . The following theorem shows that the IA does not lead to completely filled-in FOUs. ( ) UMF A 141 Theorem 5.1: Using the IA, none of the obtained FOUs will be completely filled-in, i.e., (a) for an interior FOU, b MF > a MF (see Fig. 5.9); (b) for a left-shoulder FOU, b MF ≥ a MF > 0 (see Fig. 5.10); and, (c) for a right-shoulder FOU, a MF < b MF < M (see Fig. 5.11). A proof of this theorem is given in Appendix C. C. Whose FOU? In the field of probability elicitation, O’Hagan and Oakley [34] question how various individual expert probability distributions should be combined into a single distribution, and, regardless of the method used for combining, whose distribution does this represent? The latter question reflects the fact that regardless of how the distributions are combined, the final distribution has lost the uncertainties of the individual subjects, and, in fact, it may correspond to none of the subjects. One can raise a similar question for the FOU of a word that is obtained from the IA, i.e. whose FOU does it represent? Unlike the probability elicitation field, where each expert is assumed to have a probability distribution 9 , no assumption is ever made in our work that a subject has a personal FOU for a word. An FOU is a mathematical model 9 While probability is an excellent model for unpredictable uncertainty, it is mathematics and not science; hence, to ascribe a personal probability distribution to an expert for a particular situation is fiction, regardless of how much data are elicited from that expert. A probability model may fit that data, and the fit may improve as more reliable data are extracted from the expert, but, that in no way proves that this is the expert’s personal probability distribution. 142 that captures the uncertainties about a word, and is only used in later processing or analyses. Note, however, that the union method for combining each subject’s T1 FS preserves all of their uncertainties because each of their T1 FSs is contained within the FOU. One can even provide a Bayesian-like interpretation to the construction of an FOU. Allow the analyst to be one of the subjects, so that her T1 FS is the first T1 FS of the remaining T1 FSs. Then the resulting FOU not only uses her a priori FOU, which is a T1 FS, but modifies it by folding in the T1 FSs of the remaining m m * * −1 data intervals. 5. 3 Codebook examples A data set was collected from 28 subjects at the Jet Propulsion Laboratory (JPL) 10 for a vocabulary of 32 words. These words were randomized and for all words each subject was asked the question “On a scale of 0-10, what are the end-points of an interval that you associate with the word W?”. All of the data was processed as described in Sections 5.2. Table 5.4 (in which the 32 words have been ordered using a ranking method that is explained later in this section) summarizes how many data intervals survived each of the four pre-processing stages, and how many intervals, m , survived the Deletion of Inadmissible T1 FSs step in the FS Part of the IA. Observe that m is quite variable. Table 5.4 also gives the final left- and right end-point statistics that were used to * * 10 This was done in 2002 when J. M. Mendel gave an in-house short course on fuzzy sets and systems at JPL. 143 Table 5.4: Remaining data intervals and their end-point statistics for m data intervals. Pre-processing FS Part Left-end statistic Right-end statistic Word Stage 1 n′ Stage 2 m′ Stage 3 m′′ Stage 4 m m * m l s l m r s r Teeny-weeny 28 27 25 12 12 0.0025 0.0043 0.6252 0.4478 A smidgen 26 22 20 9 9 0.0339 0.0662 0.9795 0.5869 None to very little 28 23 23 23 23 0 0 1.1371 0.3749 Tiny 28 25 25 12 12 0.0764 0.0825 1.1723 0.7466 Very little 28 23 21 13 8 0.2686 0.2205 1.2897 0.4294 Very small 28 21 21 8 6 0.3997 0.1749 1.3976 0.3137 A bit 28 25 23 12 12 1.0000 0 2.6254 0.4583 Low amount 28 26 25 17 14 0.7963 0.4278 2.6141 0.6348 Little 28 27 26 18 16 0.9612 0.2869 2.7625 0.5700 Small 28 26 25 22 20 1.2482 0.5230 3.2330 0.4653 Somewhat small 28 27 26 16 16 1.8493 0.5313 3.5449 0.5557 Some 28 26 25 17 17 2.7072 0.5834 5.2590 1.0153 Some to moderate 28 26 26 26 26 3.1236 0.6984 6.1169 0.6172 Moderate amount 28 24 23 21 21 3.8838 0.3352 6.0189 0.5770 Fair amount 28 24 22 22 22 3.9952 0.6697 6.2978 0.4932 Medium 28 24 23 20 20 4.3529 0.3853 5.8343 0.3048 Modest amount 28 21 20 19 19 4.3125 0.4369 6.0238 0.3113 Good amount 28 27 26 15 14 6.0692 0.9252 8.2282 0.4752 Sizeable 28 24 23 17 13 6.3915 0.7619 8.7298 0.7640 Quite a bit 28 27 26 14 12 6.0516 0.4770 8.3883 0.7110 Considerable amount 28 26 25 17 14 6.5049 0.7301 8.6187 0.7386 A lot 28 26 26 16 12 7.2669 0.5392 9.2706 0.4733 High amount 28 26 24 22 13 7.6442 0.7686 9.4406 0.4604 Substantial amount 28 26 25 15 10 7.6036 0.5896 9.3982 0.4499 Very sizeable 28 26 24 14 10 7.5715 0.7045 9.3577 0.4405 Large 28 17 16 16 16 7.3890 0.5366 9.0000 0 Very large 27 26 25 15 15 8.7070 0.3906 9.8906 0.1783 Humongous amount 28 23 23 22 22 9.0116 0.5658 10.0000 0 Huge amount 28 27 23 17 17 8.8236 0.3873 9.9573 0.0882 Very high amount 28 22 19 16 16 8.6886 0.5732 10.0000 0 Extreme amount 28 17 17 17 17 9.0000 0 10.0000 0 Maximum amount 27 25 23 15 15 9.2155 0.2601 10.0000 0 establish the nature of each word’s FOU. These statistics are based on the m remaining data intervals after stage 4 of pre-processing. Table 5.5 is the codebook for the 32 words. It provides the coordinates (code) for the LMF and UMF of each FOU. We warn the reader that while it may be very tempting 144 to use this codebook for your application(s), do not do this, because data collection is sensitive to scale and is application dependent; hence, although we advocate using this paper’s methodologies we do not advocate using the paper’s example word FOUs for specific applications. Table 5.5: FOU data for all words (based on m data intervals)—the Codebook. * Word LMF a UMF b Centroid Mean of centroid Teeny-weeny (0.01, 0.13) (0.14, 1.97) [0.04, 1.06] 0.550 A smidgen (0.05, 0.63) (0.26, 2.63) [0.20, 1.05] 0.625 None to very little (0.03, 0.39) (0.18, 2.63) [0.13, 1.17] 0.650 Tiny (0.05, 0.63) (0.37, 3.92) [0.21, 1.72] 0.965 Very little (0.40, 0.75, 0.41, 0.92) (0.03, 0.50, 1.25, 2.31) [0.45, 1.49] 0.970 Very small (0.40, 0.90, 0.59, 1.10) (0.19, 0.75, 1.25, 2.31) [0.66, 1.39] 1.025 A bit (0.79, 1.68, 0.74, 2.21) (0.59, 1.50, 2.00, 3.41) [1.42, 2.09] 1.755 Low amount (1.29, 1.80, 0.42, 2.21) (0.09, 1.25, 2.50, 4.62) [1.07, 3.13] 2.100 Little (1.09, 1.83, 0.53, 2.21) (0.38, 1.50, 2.50, 4.62) [1.31, 2.95] 2.130 Small (1.79, 2.28, 0.40, 2.81) (0.09, 1.50, 3.00, 4.62) [1.29, 3.34] 2.315 Somewhat small (2.29, 2.75, 0.38, 3.21) (0.59, 2.00, 3.50, 5.62) [1.73, 4.16] 2.945 Some (2.88, 3.61, 0.35, 4.21) (0.38, 2.50, 5.00, 8.00) [2.03, 5.90] 3.965 Some to moderate (4.10, 4.60, 0.36, 5.20) (0.96, 3.50, 5.50, 8.00) [2.76, 6.31] 4.535 Moderate amount (4.30, 4.75, 0.38, 5.20) (2.17, 4.00, 5.00, 7.83) [3.50, 6.28] 4.895 Fair amount (4.79, 5.29, 0.41, 6.02) (1.46, 4.25, 6.00, 8.50) [3.47, 6.78] 5.125 Medium (4.86, 5.00, 0.27, 5.14) (3.59, 4.75, 5.50, 6.90) [4.19, 6.19] 5.190 Modest amount (4.79, 5.30, 0.42, 5.70) (3.59, 4.75, 6.00, 7.41) [4.57, 6.24] 5.405 Good amount (6.79, 7.20, 0.34, 7.80) (3.17, 6.00, 8.00, 9.62) [5.04, 8.39] 6.715 Sizeable (6.79, 7.38, 0.49, 8.21) (4.38, 6.50, 8.00, 9.40) [6.16, 8.15] 7.155 Quite a bit (6.79, 7.38, 0.49, 8.21) (4.38, 6.50, 8.00, 9.40) [6.16, 8.15] 7.155 Considerable amount (7.19, 7.58, 0.37, 8.21) (4.38, 6.50, 8.25, 9.60) [5.97, 8.52] 7.245 A lot (7.69, 8.19, 0.47, 8.81) (5.38, 7.50, 8.75, 9.83) [6.99, 8.82] 7.905 High amount (7.79, 8.30, 0.53, 9.20) (5.38, 7.50, 8.75, 9.81) [7.19, 8.82] 8.005 Substantial amount (7.79, 8.30, 0.53, 9.20) (5.38, 7.50, 8.75, 9.81) [7.19, 8.82] 8.005 Very sizeable (8.29, 8.56, 0.38, 9.20) (5.38, 7.50, 9.00, 9.81) [6.94, 9.10] 8.020 Large (8.00, 8.36, 0.57, 9.17) (5.98, 7.75, 8.60, 9.52) [7.50, 8.75] 8.125 Very large (8.70, 9.90) (7.37, 9.40) [9.03, 9.58] 9.305 Humongous amount (9.74, 9.98) (7.37, 9.80) [8.71, 9.91] 9.310 Huge amount (9.47, 9.96) (7.37, 9.59) [8.88, 9.83] 9.355 Very high amount (9.34, 9.95) (7.37, 9.82) [8.97, 9.78] 9.375 Extreme amount (8.68, 9.90) (8.68, 9.90) [9.56, 9.56] 9.560 Maximum amount (9.60, 9.97) (8.68, 9.) [9.51, 9.87] 9.690 a For an interior LMF, the four numbers are (a MF , p, µ p , b MF ) ; for a left-shoulder LMF, the two numbers are (a MF , b MF ) ; and for right-shoulder LMF, the two numbers are (a MF , b MF ) . b For an interior UMF, the four numbers are a MF , C MF , C MF , b MF ) ; for a left-shoulder UMF, the two numbers are (a MF , b MF ) ; and for right-shoulder UMF, the two numbers are (a MF , b MF ) . 145 Teeny−weeny A smidgen None to very little Tiny Very little Very small A bit Low amount Little Small Somewhat small Some Some to moderate Moderate amount Fair amount Medium Modest amount Good amount Sizeable Quite a bit Considerable amount A lot High amount Substantial amount Very sizeable Large Very large Humongous amount Huge amount Very high amount Extreme amount Maximum amount Figure 5.12: FOUs for all 32 words. Start at the top row and proceed downwards scanning from left-to-right. 146 None to very little Very small A smidgen Teeny−weeny Tiny Very little A bit Small Low amount Little Somewhat small Some Some to moderate Moderate amount Fair amount Medium Modest amount Good amount Sizeable Quite a bit Considerable amount Very sizeable Substantial amount A lot High amount Very large Large Humongous amount Huge amount Very high amount Extreme amount Maximum amount Figure 5.13: FOUs for all 32 words when Stage 4 of data pre-processing is omitted. Start at the top row and proceed downwards scanning from left-to-right. 147 148 None to very little Little Moderate amount A lot Huge amount (a) Sub-vocabulary 1 Tiny Small Medium Large Very large (b) Sub-vocabulary 2 A smidgen Somewhat small Medium High amount Very high amount (c) Sub-vocabulary 3 Figure 5.14: FOUs for three sub-vocabularies. Fig. 5.12 depicts the FOUs for all 32 words. Observe that the words have been ordered so that there seems to be a very natural flow from left-shoulder FOUs to interior FOUs to right-shoulder FOUs. This flow was achieved by first computing the centroid of each FOU [20], [10], [32] and then the mean of each centroid. The results of these computations are given in the last two columns of Table 5.5. The words were then rank- ordered using the mean of the centroid. Ranking the words by this method seems to give visually acceptable results. Interestingly, sizeable and quite a bit have the same FOUs, as do high amount and substantial amount. This can be seen in Fig. 5.12, and is also evident from the LMF and UMF columns in Table 5.5. It means that based on the data collected from the 28 JPL subjects, the words sizeable and quite a bit are 100% similar, and the words high amount and substantial amount are also 100% similar. Similarity of all of the words to one another can be computed using the similarity measure for IT2 FSs that is described in [47]. Observe, also, that the word extreme amount is modeled as a T1 FS, demonstrating that the IA does provide a T1 FS model when all data intervals are the same. m * Section 5.4 argued for the inclusion of a fourth step of pre-processing, namely for reasonable-interval processing. Fig. 5.13 depicts FOUs for the 32 words when this stage was omitted. Observe that many of the FOUs are filled in or just about filled in, whereas none of the FOUs depicted in Fig. 5.12 are filled in, thereby confirming the results given in Theorem 5.1. Herrera, Herrera-Viedma and Martinez [7] discuss multi-granular linguistic term sets and how in “… decision making problems with multiple sources of information, linguistic performance values that are given to the different sources can be represented as linguistic terms sets with different granularity and/ or semantics.” Our interpretation of “linguistic term sets with different granularity” is as a sub-vocabulary from the codebook. Fig. 5.14 depicts three sub-vocabularies, where the FOUs in each sub- vocabulary cover the entire domain [0, 10]. Each sub-vocabulary was obtained from the results given in Table V and Fig. 5.12. When a codebook is established it contains within it many sub-vocabularies. One important use for a sub-vocabulary is in designing IF-THEN rules as a CWW Engine, where it is expedient to use a small (the smallest) sub-vocabulary that covers the entire domain in order to avoid rule-explosion. 149 150 5. 4 Conclusions This chapter has presented a very practical type-2-fuzzistics methodology for obtaining IT2 FS models for words, one that is called an Interval Approach (IA). The basic idea of the IA is to collect interval end-point data for a word from a group of subjects, map each subject’s data interval into a pre-specified T1 person MF interpret the latter as an embedded T1 FS of an IT2 FS, and obtain a mathematical model for the FOU for the word from these T1 FSs. The IA consists of two parts, the Data Part and the Fuzzy Set Part. In the Data Part, the interval end-point data are pre-processed, after which data statistics are computed for the surviving data intervals. In the FS Part, the data are used to decide whether the word should be modeled as an interior, left-shoulder or right-shoulder FOU. Then the parameters of the respective embedded T1 MFs are determined using the data statistics and uncertainty measures for the T1 FS models. The derived T1 MFs are aggregated using union leading to an FOU for a word, and finally to a mathematical model for the FOU. The IA has many strong points and, arguably, no weak points, i.e., it: (1) Collects interval end-point data from subjects, and this is easy to do; (2) Does not require subjects to be knowledgeable about fuzzy sets; (3) Has a straightforward mapping from data to an FOU; (4) Does not require an priori assumption about whether or not an FOU is symmetric or non-symmetric; and, (5) Leads to an IT2 FS word model that reduces to a T1 FS word model automatically if all subjects provide the same intervals. 151 Chapter 6 Conclusions and future works 6.1 Conclusions This thesis addresses two problems about water-flood management: (1) infer reservoir heterogeneity using measured injection and production rates, and (2) construct a decision support system to optimize oil production using computing with words (CWW) and the inferred reservoir heterogeneity. To infer reservoir heterogeneity, we first present an adaptive method using an Extended Kalman Filter (EKF) for the case of multiple injectors and a single producer, and then present a pseudo-virtual reservoir method for the case of multiple injectors and multiple producers, respectively. In the EKF method, a very simple parametric model, one with two parameters per injector, is used so that if a producer depends upon N injectors our model contains exactly 2N parameters. The EKF is used to adaptively estimate the 2N parameters, from which the injector-producer relationship (IPR) between each injector and a producer is then estimated. Our approach has been tested on synthetic data, reservoir simulation data and real data. Test results on synthetic data and reservoir simulation data demonstrate the feasibility of the EKF method, and test results on the real data match expert knowledge about the IPRs between injectors and a producer. All results confirm 152 that this EKF method can provide a good way to infer and track the IPRs so as to provide better insight about the IPRs. In the pseudo-virtual reservoir method, a virtual reservoir model is used to model the reservoir, and a pseudo- virtual reservoir model is used to estimate the regional impact from each injector, so that the problem for this case reduces to the problem for the case of multiple injectors and a single producer. Our approach has been tested on synthetic data, and test results suggest that this pseudo-virtual reservoir method is feasible for evaluating regional impacts with multiple injectors and multiple producers. The study for decision support system using CWW focuses on the encoder component (called type-2 fuzzistics) of a Perceptual Computer (Per-C), a specific architecture for CWW using interval type-2 fuzzy sets. Our work so far has been to transform linguistic perceptions, words, into interval type-2 fuzzy sets (IT2 FS) that activate a CWW engine. A new and simple approach, called the Interval Approach (IA) to type-2 fuzzistics has been developed; it captures the strong points of both the person membership function and interval end-points approaches. It collects interval end-point data from subjects, does not require subjects to be knowledgeable about fuzzy sets, has a straightforward mapping from data to footprint of uncertainty (FOU), does not require an a priori assumption about whether or not a FOU is symmetric or non-symmetric, and leads to an IT2 FS word model that reduces to a T1 FS word model automatically if all subjects provide the same intervals. Experiments showed that this approach is easy to implement and the derived interval type-2 word models match our intuitions, i.e., the FOUs of the small-sounding words are located to the left, the FOUs of the medium- 153 sounding words are located in the middle, and the FOUs of the large-sounding words are located to the right. 6.2 Future Works Water-flood management aims to increase secondary oil recovery. Generally, the system to fulfill such a goal operates in two stages. The first stage derives some information to characterize the reservoir using some available data, and the second stage makes decisions based on that information. This thesis adheres to this two-stage approach. Inferring reservoir heterogeneity is where we obtain the information about the reservoir heterogeneity, and constructing a decision support system is where we make better decisions based on the inferred reservoir heterogeneity and expert knowledge. To construct a sound water-flood management system, further research is needed as follows. For inferring reservoir heterogeneity, we need to perform field tests to verify our obtained test results. For constructing a decision support system using computing with words, we need to construct a decision support system using computing with words for different applications such as pattern review process and water reallocation, so that we can use the inferred reservoir heterogeneity to help us make better decisions. Some details about this proposed research are given next. 6.2.1 Perform field tests We are in the process of performing field tests on Chevron Lost Hills field data, and are ready to improve our current approach based on the feedback from the tests. 154 Based on our current research, we have found that a very important issue for current EKF and pseudo-virtual reservoir approach is validation on test results for real data. The validation for synthetic data is obvious because we know what “truth” is; however, for real data, we can’t evaluate our results by comparing estimated IPRs or regional impact with the actual IPRs or regional impact because we don’t know what the actual IPRs or regional impact are; hence, we need to use other ways to deal with this issue. Currently, our validation is to see if our results can match some expert knowledge; however, there are some limitations for such a validation procedure, namely: (1) Only a limited amount of expert knowledge is available, e.g., for Lost Hills, we only have the following two rules of expert knowledge: (a) Lost Hills is an oilfield with low permeability such that the injectors far away from the producer should impact the producer very little; and (b) there is a natural fracture around North-East 55 degree; and, (2) A reservoir is a very complex system, such that expert knowledge may only be useful for a limited number of cases, and we have no idea about when our expert knowledge is or is not useful. Hence, even with expert knowledge, we still can’t obtain convincible validation on test results because we can’t tell if the validation is correct. Consequently, we need to design a new approach to deal with this. One approach is to forecast the production rates of producers using our EKF after a deliberate change has been made to an injector. By comparing the forecasted production rates with the actual measured production rates, we may gain more confidence on our 155 test results and be more confident of the validation. Two issues need to be examined: 1) Develop a strategy to change the injection rates. From the view of Chevron engineers, changing the injection rate may change the reservoir model, which directly relates to our estimated IPRs or regional impacts; hence, how to change the injection rate such that the reservoir model will not change much is an important issue that needs to be examined. 2) Develop a procedure for validation. Because the reservoir is a dynamic system, even if we don’t intentionally change anything, the reservoir model may change constantly and naturally; hence, we also need to develop a procedure for validation such that it can correctly describe our test results, e.g., we may validate the test results in a small amount of time so that we can assume that the reservoir model does not change too much in that small amount of time. 6.2.2 Construct a Decision Support System Using Computing With Words After inferring the reservoir heterogeneity, the next question we would like to address is what we can do for inferred reservoir heterogeneity. Pattern review process could be one such application that uses the inferred reservoir heterogeneity to make better decisions. As an important part of water-flood management, the pattern review process often leads to the determination of remediation steps to optimize recovery and mitigate subsidence. Currently, the pattern review is performed by human experts using geology information (such as location, fault information, etc.) and production rates, etc. However, there are some limitations to this: 156 1) Because of the enormous amount of available data and the very complex reservoir environment, this process is very difficult and somewhat inefficient; 2) Neat patterns defined by geology information can’t actually determine the correct relationships between injectors and producers, i.e., petroleum engineers know that the injectors that relate to a particular producer are frequently not always within the pattern of this producer. Sometimes, they are two or three patterns away. Sometimes, they are even farther away, bypassing some close producers and going to a more distant producer. This is more prevalent in highly fractured reservoirs. Consequently, our long-term goal is to provide a decision support tool to supplement Chevron engineers carrying out asset activities and processes, such that a pattern can be identified automatically. The pattern would be identified based not only on geological information and other Chevron expert knowledge, but also on inferred reservoir heterogeneity, where the latter would be obtained using our newly developed EKF and pseudo-virtual reservoir approaches. Future work should be conducted to develop a decision support CWW system for the pattern review process, because the CWW using interval type-2 FLS is very suitable when: o Knowledge is mined from experts using IF-THEN questionnaires, but there is uncertainty about the meanings of the questionnaire words, because words mean different things to different people; and, o Computation is done with words (words-in and words-out). Some issues that should be studied are: • Design a way to extract some expert knowledge about the pattern review process from Chevron Engineers. Usually, the expert knowledge is captured in the form of rules, which are mined from experts using IF-THEN questionnaires. An example of a rule is: : IF the - is , and the between producer and injector is , THEN this injector and producer are within the same p l R injector producer relation low fracture highly impossible very unlikely to be attern. (6.1) In (6.1), rule antecedents (the IF parts of a rule) injector-producer relation and fracture involve the forecasted injector-producer relationships using our newly developed Extended Kalman Filtering approach, and the geology information about the fracture; and each rule consequent (the THEN part of a rule), e.g., very unlikely, is a term to describe the pattern information. Each expert may provide some rules, and then for a group of experts, a group of rules can be obtained to support decision making. • Develop and implement a strategy to construct the CWW decision support system. When we construct the CWW decision support system, we need to decide if the system is distributed and hierarchical. • Test this decision support system using real data and an interaction with Chevron engineers. This is very important for validating this decision system. 157 6.2.3 Some future works about interval approach for type-2 fuzzistics Recall from Chapter 5 that, so far, it has been assumed that interval end-point data have been collected from a group of subjects. Although we strongly advocate doing this, so that the data will contain both intra- and inter-levels of uncertainty, we realize that there may be times when this is not possible, due, for example, to budget limitations, time-constraints, or unavailability of a subject pool. How to obtain an FOU from a single subject by an IA is an open research issue. Finally, the following validation problem needs to be studied. Given an FOU whose parameters are specified, use it to generate N symmetrical triangle embedded T1 FSs. Such T1 FSs might be called the “realizations” of the FOU T1 FS generator. Using the equations that are given in Table 5.3, but in reverse, obtain ( for each of the N . Prove that by using the IA, one can obtain the original FOU to within a quantified level of accuracy (that will depend upon N), i.e., prove some sort of convergence result for the IA. Because this validation problem is very different from the problem that is considered in Chapter 5, it will appear as the future work for type-2 fuzzistics. a (i) ,b (i) ) (a MF (i) ,b MF (i) ) 158 APPENDIX A THE MODEL OF INJECTION RATES To test our proposed approach, we need to design the reasonable model, which matches practical situations, for injection rate data generation. Our understanding of the practical case is follows: when the engineer injects water at time t, he firstly has the scheduled set point of injection rate, called scheduled injection rate, which can be the same as or different from that at different time t ′. However, because of all kind of issues, the actual injection rate has some deviation from the scheduled injection rate, e.g., 30%, 50% deviation. Therefore, the actual injection rate is different with the scheduled injection rate. Additionally, the environment may introduce some noises into the measurement of injection rate, such that the measured injection rate composes of the actual injection rate and measurement noises. Based on this understanding, the simulation model of injection rate generation for the injector is: th i ( ) () 1 ( ) ( ) is i i InI Vn Nn =+ + (A-1) where () i I n , s I are the measured injection rate and scheduled injection rate, respectively, is the zero-mean random deviation about the set point, () i Vn ( ) 1 () s i I Vn + is the actual injection rate, and is the measurement noise of injection rate. ( ) i Nn As some experts said, usually, the s I is fixed, but there are infrequent changes; and the is about 5%-7% of injection rate. Normally, we assume that the measurement () i Nn 159 noise is zero-mean Gaussian distributed, and independent on the actual injection rate () i Nn ( ) 1() s i I Vn + . Also, we assume that the is zero-mean uniform distributed random variable in [ ( ) i Vn ] , a a − . Based on these assumptions, we can obtain the mean and variance of ( ) i I n, and as: () i V n ( ) i N n 22 222 0 1 3 1 3 Is VN V Is mI mm a Ia δ 2 N δ δ = ⎧ ⎪ == ⎪ ⎪ ⎨ = ⎪ ⎪ =+ ⎪ ⎩ (A-2) where , , are the mean of I m V m N m ( ) i I n, and , respectively, and () i V n ( ) i N n 2 I δ , 2 V δ and 2 N δ are the variance of ( ) i I n, and , respectively. () i V n ( ) i N n The derivation of (A-2) is follows. As we have assumed, the is zero-mean Gaussian distributed, and independent on the actual injection rate ( ) i N n ( ) 1 () s i I Vn + , and the is zero-mean uniform distributed random variable in ( ) i Vn [ ] , aa − , it follows that 0 VN mm = = (A-3) Hence, ( ) { } 1() () I si i mEI Vn Nn I s = ++ = (A-4) 160 Let 2 N δ be the variance of , and the variance of , which is uniformly distributed in [ ( ) i Nn ( ) i Vn ] , a a − , can be computed by: 22 11 (0) 23 a V a 2 x dx a a δ − =− = ∫ (A-5) Because is independent on the actual injection rate () i N n ( ) 1 () s i I Vn + , the variance of () i I n is () { } ( ) { } { } 2 22 2 2 2 2 var 1 () () var 1 () var ( ) 1 3 Is i i s i i sV N s N I Vn N n I V n N n IIa δ δδ δ =+ + = + + =+ = + (A-6) which completes the derivation of (A-2). To generate the simulation data, we need to make choices on ( ) i I n, and . To make the choice more practically, we refer our choices to the mean and variance of a collection of injection rates from 10 injectors, which are summarized in Table A-1. () i Vn ( ) i N n 161 As we had discussed, let’s assume 22 0.05 NI δ δ ≈ , then 22 2 1 0.95 3 s I Ia δ ≈ . Hence, (A-7) () () 2 2 2.85 / 1.6882 / Is Is aI aI δ δ ⎧ ≈ ⎪ ⎪ ⎨ ⎪ ≈ ⎪ ⎩ Table A-1: the mean and variance of the injection rates from 10 injectors. Name of injector Mean s I Variance 2 I δ Is I δ I5 12- 1B 2136 508369 713/2136 = 0.3338 I5 11- 1W 2465 269361 519/2465 = 0.2105 I5 11- 1WA 2707 527076 727/2707 = 0.2686 I5 10- 1W 2392 373321 611/2392 = 0.2554 I5 10- 2WAS 1814 148225 385/1814 = 0.2122 I5 11-2WB (Before Aug. 2002) 2658 318096 564/2658 = 0.2122 I5 10- 2WB 1577 91204 302/1577 = 0.1915 I5 11- 2W 2341 320356 566/2341 = 0.2418 I5 11- 2WAL 2332 275625 525/2332 = 0.2251 I5 9- 2W 2522 625681 791/2522 = 0.3136 From Table A-1, we can see that the I s I δ is about 0.19 – 0.33. Therefore, ( ) 1.6882 / 0.3208 0.5571 Is aI δ ≈≈ ∼ (A-8) which means the deviation of the scheduled injection rate is about 32% to 56%. 162 In summary, my choices on these variables are: (1) . 2000 s I = (2) such that is the uniform random variable in domain [ 0.4 a = ( ) i Vn ] 0.4, 0.4 − . (3) 22 22 2 1 (2000 0.4) 3 5% 106 95% 3 19 3 19 s s N Ia Ia δ × =×= = = ×× . 163 APPENDIX B DERIVATION OF REASONABLE INTERVAL TEST FOR (5.4)-(5.6) In this Appendix, derivations of (5.4)-(5.6) are obtained. Examining Fig. 5.4, and using the requirement that reasonable data intervals must overlap, it must be true that: (B-1) min ∀i =1,..., ′′ m b (i) > max ∀i =1,..., ′′ m a (i) A simple way to satisfy (B-1) is to require that: a (B-2) (i) < ξ b (i) > ξ ⎫ ⎬ ⎭ , ∀i = 1,..., ′′ m where threshold ξ has to be chosen, and there can be different ways to do this. In this paper, an optimal value of ξ , ξ* , is chosen so that (B-3) ξ* = arg min[ ξ P(a (i) > ξ) + P(b (i) < ξ)] By choosing ξ * in this way, data intervals that don’t satisfy (B-2) will occur with the smallest probability 11 . 11 Another approach might be to choose ξ * that maximizes . As of yet, we have not done this. P[a (i) < ξ,b (i) > ξ] 164 In order to compute ξ*, it is assumed that each a ( (i) i = 1,..., ′′ m ) is Gaussian with mean and standard deviation m a σ a , and each ( b (i) i = 1,..., ′ ′ m ) is also Gaussian, but with mean and standard deviation m b σ b . It follows that: P(a (i) > ξ) +P(b (i) < ξ) = 1 2 π σ a e − 1 2 a (i ) −m a σ a ⎡ ⎣ ⎢ ⎢ ⎤ ⎦ ⎥ ⎥ 2 da (i) ξ ∞ ∫ + 1 2 π σ b e − 1 2 b (i ) −m b σ b ⎡ ⎣ ⎢ ⎢ ⎤ ⎦ ⎥ ⎥ 2 db (i) −∞ ξ ∫ (B-4) Setting the derivative of this function with respect to ξ equal to zero, ξ* is found to be the solution of: 1 2 π σ a e − 1 2 ξ* −m a σ a ⎡ ⎣ ⎢ ⎤ ⎦ ⎥ 2 = 1 2 π σ b e − 1 2 ξ* −m b σ b ⎡ ⎣ ⎢ ⎤ ⎦ ⎥ 2 (B-5) Observe that ξ * occurs at the intersection of the two Gaussian distributions and . Taking the natural logarithm of both sides of (B-5), one is led to the following quadratic equation: p(a (i) ) p(b (i) ) ( σ a 2 − σ b 2 ) ξ * 2 +2(m a σ b 2 −m b σ a 2 ) ξ * +[m b 2 σ a 2 −m a 2 σ b 2 − 2 σ a 2 σ b 2 ln( σ a / σ b )] =0 (B-6) 165 The two solutions of this equation are: ξ* = (m b σ a 2 −m a σ b 2 ) ± σ a σ b (m a −m b ) 2 + 2( σ a 2 − σ b 2 )ln( σ a / σ b ) ⎡ ⎣ ⎤ ⎦ 1/2 ( σ a 2 − σ b 2 ) (B-7) The final solution is chosen as the one for which: ξ* ∈[m a ,m b ] (B-8) That this solution minimizes P , rather than maximizes it follows from showing that the derivative of (B-5) with respect to (a (i) > ξ) +P(b (i) < ξ) ξ , after which ξ is set equal to ξ*, is positive. Because this is a very tedious calculation, an alternative is presented next. Fact: is a concave function and its minimum value, P(a (i) > ξ) +P(b (i) < ξ) ξ * , occurs in the interval [ . m a ,m b ] A proof of this fact follows from (a) and (b) of Fig. B-1. From Fig. B-1 (a), observe that at ξ = ξ *, (B-9) P(a (i) > ξ*) +P(b (i) < ξ*) = A 1 + (A 2 + A 3 ) 166 and at ξ = ′ ξ ; (B-10) hence, P(a (i) > ′ ξ ) +P(b (i) < ′ ξ ) = (A 1 + A 2 + A 4 ) + A 3 (B-11) P(a (i) > ′ ξ ) +P(b (i) < ′ ξ ) > P(a (i) > ξ*) +P(b (i) < ξ*) Proceeding in a similar manner for Fig. B-1 (b), it follows that: (B-12) P(a (i) > ′′ ξ ) +P(b (i) < ′′ ξ ) > P(a (i) > ξ*) +P(b (i) < ξ*) (B-11) and (B-12) together prove that P is a concave function about (a (i) > ξ) +P(b (i) < ξ) ξ = ξ *. The concave shape of is depicted in Fig. B-1 (c). P(a (i) > ξ) +P(b (i) < ξ) Above, it has been proven that ξ * occurs at the intersection of and ; but, it is clear from Fig. 5.B-1 (a) or (b) p(a (i) ) p(b (i) ) ξ* ∈[m a ,m b ]. Q. E. D. Because access to the population means and standard deviations is unavailable, they must be estimated in order to compute ξ* and to perform the test in (B-8). Our approach is to estimate those quantities as: (B-13) ˆ m a = m l ˆ m b = m r ˆ σ a = σ l ˆ σ b = σ r ⎫ ⎬ ⎪ ⎪ ⎭ ⎪ ⎪ 167 Doing this, one obtains (5.4)-(5.6). Note that numerical values for m , , l m r σ l and σ r are available at the end of Tolerance Limit Processing, so that ξ* can indeed be computed. a m b m ( ) p • () () i pa 2 A () () i pb 4 A 168 (a) (b) (c) Figure B-1: , and the four areas (i p(a (i) ) p(b (i) ) A i = 1,..., 4 ) that can be used to compute , and (a) , or (b) ; and, (c) the concave shape of . P(a (i) > ξ*) +P(b (i) < ξ*) P(a (i) > ′ ξ ) +P(b (i) < ′ ξ ) P(a (i) > ′′ ξ ) +P(b (i) < ′′ ξ ) P(a (i) > ξ) +P(b (i) < ξ) a m b m ( ) p • * ξ ξ ′ ′ () () i pa () () or ii ab () () i pb 3 A 2 A 4 A 1 A a m b m * ξ ξ ′ ′ ξ ′ () ( ) ()( ii Pa Pb ) ξ ξ >+ < ξ 1 * ξ ξ ′ 1 A 3 A () () or ii ab APPENDIX C PROOF OF THEOREM 5.1 This Appendix contains proofs for the three parts of Theorem 5.1. It uses a MF , a MF , b MF and b MF that are defined in (22) and (23). (a) Using the equations for a and b , that are given in the top row of Table 5.3, it follows that ( ∀ ) MF (i) MF (i) i = 1,...,m* a MF (i) = 1 2 [(a (i) +b (i) ) − 2(b (i) −a (i) )] = a (i) − 2 − 1 2 (b (i) −a (i) ) <a (i) (C-1) b MF (i) = 1 2 [(a (i) +b (i) ) + 2(b (i) −a (i) )] =b (i) + 2 − 1 2 (b (i) −a (i) ) >b (i) (C-2) Let a ≡ max i =1,...,m* {a (i) } (C-3) b ≡ min i =1,...,m* {b (i) } (C-4) Applying the definitions of a MF and b MF as well as (C-3) and (C-4) to (C-1) and (C- 2), it follows that: 169 a MF <a (C-5) b MF > b (C-6) From (5.4), (C-3) and (C-4), it is true that: a < ξ* < b (C-7) Substituting (C-5) and (C-6) into (C-7), it follows that: a MF < ξ* < b MF (C-8) Consequently, b MF >a MF (C-9) (b) Using the equation for , given in the second row of Table 5.3, it follows that: a MF (i) a MF (i) = (a (i) +b (i) ) 2 − (b (i) −a (i) ) 6 = 1 2 + 1 6 ⎛ ⎝ ⎜ ⎞ ⎠ ⎟ a (i) + 1 2 − 1 6 ⎛ ⎝ ⎜ ⎞ ⎠ ⎟ b (i) > 0 (C-10) 170 Applying the definition of a MF to (C-10), it follows that: a MF >0 (C-11) Additionally, it is always true that b ; (C-12) MF (i) ≥a MF (i) ∀i = 1,...,m* hence, applying the definitions of b MF and a MF to (C-12), it follows that: b MF ≥ a MF (C-13) Combining (C-13) and (C-11), it follows that b MF ≥ a MF > 0 ; hence, b MF > a MF (C-14) (c) Using the equation for b , given in the third row of Table 5.3, it follows that: MF (i) b MF (i) = M − ( ′ a (i) + ′ b (i) ) 2 + ( ′ b (i) − ′ a (i) ) 6 = (b (i) +a (i) ) 2 + (b (i) −a (i) ) 6 (C-15) 171 Consequently, b MF (i) < (a (i) +b (i) ) 2 + (b (i) −a (i) ) 2 =b (i) (C-16) Applying the definitions of b MF and b to (C-16), it follows that: b MF <b; (C-17) but, it is also true that: b (C-18) (i) ≤ M ∀i = 1,...,m * so that b ≤ M (C-19) Combining (C-17) and (C-19), observe that: b MF < M (C-20) 172 Next, using the equation for , given in the third row of Table 5.3, it follows that: a MF (i) a MF (i) = M − ( ′ a (i) + ′ b (i) ) 2 − 6( ′ b (i) − ′ a (i) ) 3 = (b (i) +a (i) ) 2 − 6(b (i) −a (i) ) 3 (C-21) Comparing (C-21) and (C-15), observe that: b (C-22) MF (i) >a MF (i) ∀i = 1,...,m* Applying the definitions of b MF and a MF to (C-22), it follows that: b MF >a MF (C-23) Finally, combining (C-20) and (C-23), one obtains: a MF <b MF < M (C-24) 173 174 References [1] Alejandro Albertoni and Larry W. Lake, “Inferring Interwell Connectivity only from Well-Rate Fluctuations in Waterfloods,” SPE 83381, 2003. [2] A. N. Araqye-Martinez, “Estimation of autocorrelation and its use in sweep efficiency calculation,” MS thesis, U. of Texas at Austin, Austin, Texas, 1993. [3] D. Dubois, L. Foulloy, G. Mauris and H. Prade, “Probability-possibility transformations, triangular fuzzy sets, and probabilistic inequalities,” Reliable Computing, vol. 10, pp. 273-297, 2004. [4] C. Eisenhart, M. W. Hastay and W. A. Wallis, Techniques of Statistical Analysis, Ch. 2, McGraw-Hill Book Co. NY, 1947. [5] Simon Haykin, Kalman Filtering and Neural Networks. John Wiley & Sons, 2001. [6] Heffer, K. J., Fox, R. J., and McGill, C.A., “Novel Techniques Show Links Between Reservior Flow Directionality, Earth Stree, Fault Structure and Geomechanical Changes in Mature Waterfloods,” Paper SPE30711 Presented at the 1995 SPE Annual Technical Conference and Exhibition, Dallas, 22-25, Oct., 1995. [7] F. Herrera, E. Herrera-Viedma and L. Martinez, “A fusion approach for managing multi-granularity linguistic term sets in decision making,” Fuzzy Sets and Systems, vol. 114, pp. 43-58, 2000. [8] F. Herrera and L. Martinez, “A type-2 fuzzy linguistic representation model for computing with words,” IEEE Trans. on Fuzzy Systems, vol. 8, no. 6, pp. 746- 752, 2000. [9] J. Kacprzyk and R.R. Yager, ”Linguistic summaries of data using fuzzy logic,” Int’l. J. of General Systems, vol. 30, pp. 33-154, 2001. [10] N. N. Karnik and J. M. Mendel, “Centroid of a type-2 fuzzy set,” Information Sciences, vol. 132, pp. 195-220, 2001. 175 [11] George. J. Klir, “Uncertainty and Information: Foundations of Generalized Information Theory”, John Wiley & Sons, Inc., New York, 2006. [12] J. Lawry, “A methodology for computing with words,” Int’l. J. of Approximate Reasoning, vol. 28, pp. 51-89, 2001. [13] Feilong Liu and Jerry M. Mendel, “Aggregation Using Fuzzy Weighted Averages, as computed by the Karnik-Mendel Algorithm”, Accepted by IEEE Transaction on Fuzzy Systems, 2006 [14] Feilong Liu and Jerry M. Mendel, “On Computing the Fuzzy Weighted Average Using the KM Algorithm”, Proceeding of IPMU 2006, Paris, France, 2006. [15] Feilong Liu, “An Efficient Centroid Type Reduction Strategy for General Type- 2 Fuzzy Logic System”, The website of IEEE Computational Intelligence Society (http://ieee-cis.org/edu/research/), 2006. [16] Feilong Liu and Jerry M. Mendel, “An interval approach to Fuzzistics for interval type-2 fuzzy sets”, Accepted by FUZZ-IEEE 2007, London, UK, 2007. [17] Feilong Liu and Jerry M. Mendel, “Forecasting inter-well relation by using only injection rates and production rates”, Accepted by ATCE2007, 2007. [18] M. Margaliot and G. Langholz, “Fuzzy control of a benchmark problem: A computing with words approach,” IEEE Trans. on Fuzzy Systems, vol.12, no. 2, pp. 230-235, 2004. [19] Jerry M. Mendel, Lessons in Estimation Theory for Signal Processing, Communications, and Control. Prentice Hall PTR, 1995. [20] J. M. Mendel, “Computing with words, when words can mean different things to different people,” in Proc. 3rd International ICSC Symposium on Fuzzy Logic and Applications, Rochester Univ., Rochester, NY, June 1999, pp. 158-164. [21] J. M. Mendel, Uncertain Rule-Based Fuzzy Logic Systems: Introduction and New Directions, Prentice-Hall, Upper Saddle River, NJ, 2001. [22] J. M. Mendel and R. I. Bob John, “Type-2 fuzzy sets made simple,” IEEE Trans. on Fuzzy Systems, vol. 10, pp. 117-127, April 2002. 176 [23] J. M. Mendel, “An architecture for making judgments using computing with words,” Int’l. J. of Applied Mathematics and Computer Science, Vol. 12, no. 3, pp. 325-335, 2002. [24] Jerry M. Mendel, Robert I. John and Feilong Liu, “On Using Type-2 Fuzzy Set Mathematics to Derive Interval Type-2 Fuzzy Logic System”, NAFIPS’2005, Ann Arbor, Michigan, 2005. [25] Jerry M. Mendel and Feilong Liu, “Super-Exponential Convergence of the KM Algorithms that are used for Type-Reduction in a Type-2 Fuzzy Logic System”, Proceeding of IEEE WCCI 2006, Vancouver, BC, Canada, 2006. [26] J. M. Mendel and H. Wu, “Type-2 fuzzistics for symmetric interval type-2 fuzzy sets: Part 1: forward problem,” IEEE Trans. on Fuzzy Systems, Vol. 14, pp. 781-792, 2006. [27] J. M. Mendel and H. Wu, “Type-2 fuzzistics for symmetric interval type-2 fuzzy sets: Part 2: Inverse problem,” IEEE Trans. on Fuzzy Systems, 2007, in press. [28] J. M. Mendel, “On eliciting membership function information from subjects,” submitted for publication, 2006. [29] Jerry M. Mendel, Robert I. John and Feilong Liu, “Interval Type-2 Fuzzy Logic Systems Made Simple”, IEEE Transaction on Fuzzy Systems, vol. 14, pp. 808- 821, No. 6, Dec., 2006. [30] Jerry M. Mendel and Feilong Liu, “Super-Exponential Convergence of the Karnik-Mendel Algorithm for Computing the Centroid of an Interval Type-2 Fuzzy set”, IEEE Transaction on Fuzzy Systems, vol. 15, No. 2, pp. 309-322, April, 2007. [31] J. M. Mendel, “Computing with words and its relationship with fuzzistics,” Information Sciences, Vol.177, pp. 988-1006, 2007. [32] J. M. Mendel and H. Wu, “New results about the centroid of an interval type-2 fuzzy set, including the centroid of a fuzzy granule,” Information Sciences, vol. 177, pp.360-377, 2007. 177 [33] J. M. Mendel, “Type-2 fuzzy sets and systems: an overview,” IEEE Computational Intelligence Magazine, vol. 2, pp. 20-29, February 2007. [34] A. O’Hagan and J. E. Oakley, “Probability is perfect, but we can’t elicit it perfectly,” Reliability Engineering and System Safety, vol. 85, pp. 239-248, 2004. [35] M. N. Panda and A. K. Chopra, “An integrated approach to estimate well interaction,” paper SPE 39563 presented at the 1998 SPE India oil and gas conference and exhibition, New Delhi, Feb., 17-19, 1998. [36] De Sant’Anna Pizarro, J. O., “Estimating injectivity and lateral autocorrelation in heterogeneous media,” Ph.D dissertation, U. of Texas at Austin, Austin, Texas, 1998. [37] B. T. Refunjol, “Reservoir characterization of North Buck Draw field based on tracer response and production/injection analysis,” MS thesis, U. of Texas at Austin, Austin, Texas, 1996. [38] S. H. Rubin, “Computing with words,” IEEE Trans. on Systems, Man and Cybernetics, Part B, vol. 29, no. 4, pp. 518-524, 1999. [39] T. Soeriawinata and M. Kelkar, “Reservoir management using production data,” paper SPE 52224 presented at the 1999 SPE Mid-Continent operation symposium, Oklahoma City, Oklahoma, March 28-31, 1999. [40] Thakur, Ganesh C. and Abdus Satter, Integrated Waterflood Asset Management. PennWell Corporation, 1998. [41] I. B. Turksen, “Type-2 representation and reasoning for CWW,” Fuzzy Sets and Systems, vol. 127, pp. 17-36, 2002. [42] Stephen B. Vardeman, “Statistics for Engineering Problem Solving”, IEEE Press, New York, 1994. [43] R. W. Walpole, R. H. Myers, A.L. Myers and K. Ye, Probability & Statistics for Engineers and Scientists,” 8 th Edition, Prentice-Hall, Upper Saddleback River, NJ, 2007. 178 [44] H. Wang and D. Qiu, “Computing with words via Turing machines: A formal approach,” IEEE Trans. on Fuzzy Systems, vol.11, no. 6 , pp. 742-753, 2003. [45] J. H. Wang and J. Hao, “A new version of 2-tuple fuzzy linguistic representation model for computing with words,” IEEE Trans. on Fuzzy Systems, vol. 14, no. 3, pp. 435-445, 2006. [46] D. Wu and J. M. Mendel, “The linguistic weighted average,” Proc. of 2006 IEEE Int’l. Conf. on Fuzzy Systems, pp. 3030-3037, Vancouver, CA, July 2006. [47] D. Wu and J. M. Mendel, “A vector similarity measure for interval type-2 fuzzy sets and type-1 fuzzy sets,” submitted for publication, 2006. [48] R. R. Yager, “On the retranslation process in Zadeh’s paradigm of computing with words,” IEEE Trans. on Systems, Man and Cybernetics, Part B, vol.34, no. 2, pp. 1184-1195, 2004. [49] R. R. Yager, “A new approach to the summarization of data,” Information Sciences, vol. 28, pp. 69-86, 1982. [50] A. A. Yousef, P. Gentil, J. L. Jensen and L. W. Lake, “A Capacity Model to Infer Interwell Connectivity From Production and Injection Rate Fluctuations,” SPE 95322, Presented at the SPE ATCE 2005, Dallas, Oct., 2005. [51] L. A. Zadeh, “Fuzzy Logic = Computing with words,” IEEE Trans. on Fuzzy Systems, vol. 4, pp. 103-111, 1996. [52] L. A. Zadeh, “From computing with numbers to computing with words—from manipulation of measurements to manipulation of perceptions,” IEEE Trans. on Circuits and Systems–I: Fundamental Theory and Applications, vol. 4, pp. 105- 119, 1999.
Abstract (if available)
Abstract
This thesis addresses two problems about water-flood management: (1) infer reservoir heterogeneity using measured injection and production rates, and (2) construct a decision support system to optimize oil production using computing with words (CWW) and the inferred reservoir heterogeneity.
Linked assets
University of Southern California Dissertations and Theses
Conceptually similar
PDF
Investigating statistical modeling approaches for reservoir characterization in waterfloods from rates fluctuations
PDF
Real-time reservoir characterization and optimization during immiscible displacement processes
PDF
Aggregation and modeling using computational intelligence techniques
PDF
Stochastic oilfield optimization under uncertainty in future development plans
PDF
Advances in linguistic data-oriented uncertainty modeling, reasoning, and intelligent decision making
PDF
Parameter estimation to infer injector-producer relationships in oil fields: from hybrid constrained nonlinear optimization to inference in probabilistic graphical model
PDF
Intelligent systems for decision support
PDF
A method for characterizing reservoir heterogeneity using continuous measurement of tracer data
PDF
Waterflood induced formation particle transport and evolution of thief zones in unconsolidated geologic layers
PDF
Modeling and simulation of complex recovery processes
PDF
A coupled multiphase flow-geomechanics framework for modeling material, kinematic and contact nonlinearities in faulted reservoirs
PDF
Optimization of coupled CO₂ sequestration and enhanced oil recovery
PDF
History matching for fractured reservoirs with Hough-transform-based parameterization
PDF
Mass transfer during enhanced hydrocarbon recovery by gas injection processes
PDF
Coarse-scale simulation of heterogeneous reservoirs and multi-fractured horizontal wells
PDF
Flood front tracking and continuous recording of time lag in immiscible displacement
PDF
Integration of multi-physics data into dynamic reservoir model with ensemble Kalman filter
PDF
Reactivation of multiple faults in oilfields with injection and production
PDF
Learning and control for wireless networks via graph signal processing
PDF
Intelligent knowledge acquisition systems: from descriptive to predictive models
Asset Metadata
Creator
Liu, Feilong
(author)
Core Title
Intelligent signal processing for oilfield waterflood management
School
Viterbi School of Engineering
Degree
Doctor of Philosophy
Degree Program
Electrical Engineering
Publication Date
04/11/2008
Defense Date
03/14/2008
Publisher
University of Southern California
(original),
University of Southern California. Libraries
(digital)
Tag
computing with words,decision support system,extended Kalman filter,fuzzistics,injector-producer relationship,interval approach,OAI-PMH Harvest,oilfield waterflood management,regional impact,reservoir connectivity
Language
English
Advisor
Mendel, Jerry M. (
committee chair
), Ershaghi, Iraj (
committee member
), Jenkins, B. Keith (
committee member
)
Creator Email
feilongl@usc.edu
Permanent Link (DOI)
https://doi.org/10.25549/usctheses-m1102
Unique identifier
UC1178185
Identifier
etd-Liu-20080411 (filename),usctheses-m40 (legacy collection record id),usctheses-c127-54766 (legacy record id),usctheses-m1102 (legacy record id)
Legacy Identifier
etd-Liu-20080411.pdf
Dmrecord
54766
Document Type
Dissertation
Rights
Liu, Feilong
Type
texts
Source
University of Southern California
(contributing entity),
University of Southern California Dissertations and Theses
(collection)
Repository Name
Libraries, University of Southern California
Repository Location
Los Angeles, California
Repository Email
cisadmin@lib.usc.edu
Tags
computing with words
decision support system
extended Kalman filter
fuzzistics
injector-producer relationship
interval approach
oilfield waterflood management
regional impact
reservoir connectivity