作者: LATE 時(shí)間: 2025-3-21 23:33 作者: follicular-unit 時(shí)間: 2025-3-22 02:22
Numerical Methods for Computing Stationary Distributions of Finite Irreducible Markov Chains,let .. denote the rate at which an .-state Markov chain moves from state . to state .. The . × . matrix . whose off-diagonal elements are .. and whose .. diagonal element is given by ... is called the . of the Markov chain. It may be shown that the stationary probability vector ., a row vector whose作者: 缺乏 時(shí)間: 2025-3-22 06:00
Stochastic Automata Networks,is represented by a number of states and rules that govern the manner in which it moves from one state to the next. The state of an automaton at any time . is just the state it occupies at time . and the state of the SAN at time . is given by the state of each of its constituent automata.作者: ovation 時(shí)間: 2025-3-22 09:26 作者: 綁架 時(shí)間: 2025-3-22 14:56
Use of Characteristic Roots for Solving Infinite State Markov Chains,pace. In particular, we consider the solution of such chains using roots or zeros. A root of an equation . (.) = 0 is a zero of the function . (.),and so for notational convenience we use the terms root and zero interchangeably. A natural class of chains that can be solved using roots are those with作者: 綁架 時(shí)間: 2025-3-22 18:58 作者: Arctic 時(shí)間: 2025-3-23 00:47 作者: Lucubrate 時(shí)間: 2025-3-23 03:20
On Numerical Computations of Some Discrete-Time Queues,As-mussel, 1987, Bacelli and Bremaud, 1994, Bhat and Basawa, 1992, Boxma and Syski, 1988, Bunday, 1986, Bunday, 1996, Chaudhry and Templeton, 1983, Cohen, 1982, Cooper, 1981, Daigle, 1992, Gnedenko and Kovalenko, 1989, Gross and Harris, 1985, Kalashnikov, 1994, Kashyap and Chaudhry, 1988, Kleinrock,作者: dyspareunia 時(shí)間: 2025-3-23 06:39 作者: AND 時(shí)間: 2025-3-23 10:53
Techniques for System Dependability Evaluation,computer and communication systems. While system . has received a lot of attention in the past, increasingly system . is gaining importance. The proliferation of computer and computer-based communication systems has contributed to this in no small measure. This chapter is thus a step in the directio作者: Flavouring 時(shí)間: 2025-3-23 15:42
Robert Handfield,Daniel J. FinkenstadtMany man-made systems, especially those in the areas of computer and communication, are so complex that it is essential to study them with simplified mathematical models during their design, prototyping, and deployment.作者: 孤獨(dú)無助 時(shí)間: 2025-3-23 20:12
Tools for Formulating Markov Models,Many man-made systems, especially those in the areas of computer and communication, are so complex that it is essential to study them with simplified mathematical models during their design, prototyping, and deployment.作者: Indebted 時(shí)間: 2025-3-23 22:16 作者: Scintigraphy 時(shí)間: 2025-3-24 04:29
https://doi.org/10.1007/978-3-658-02664-6n interval (0, .) is “sufficiently large” (. → ∞). These measures are indeed approximations of the behavior of the system for a finite, but long, time interval, where long means with respect to the interval of time between occurrences of events in the system. However, an increasing number of applica作者: gangrene 時(shí)間: 2025-3-24 07:29
https://doi.org/10.1007/978-3-658-02664-6let .. denote the rate at which an .-state Markov chain moves from state . to state .. The . × . matrix . whose off-diagonal elements are .. and whose .. diagonal element is given by ... is called the . of the Markov chain. It may be shown that the stationary probability vector ., a row vector whose作者: Gum-Disease 時(shí)間: 2025-3-24 14:32 作者: vanquish 時(shí)間: 2025-3-24 15:07
Fallstudien Supply Chain Integration,GI/M/1-type processes are Markov chains with transition matrices having the same structure as the imbedded Markov chain of a GI/M/1 queue, except that the entries are matrices rather than scalars. Similarly, M/G/1 type processes have transition matrices of the same form as the imbedded Markov chain 作者: 個(gè)阿姨勾引你 時(shí)間: 2025-3-24 20:16 作者: 急急忙忙 時(shí)間: 2025-3-25 00:09 作者: ADORN 時(shí)間: 2025-3-25 04:51
Helmuth Ludwig,Alastair Orchardchain to achieve an economic objective. Control is exercised by taking a sequence of actions, each of which may depend on the currently observed state and may influence both the immediate cost and the next state transition.作者: Dissonance 時(shí)間: 2025-3-25 10:44 作者: 鴿子 時(shí)間: 2025-3-25 13:11
Disruptive Innovation Through 3D Printing of the research effort has been devoted to so-called Jackson networks, that is, networks with Poisson arrivals, exponential service times and routing independent of the state of the system and the history of the customer. The steady-state distribution of Jackson networks can be expressed in a so-ca作者: 含鐵 時(shí)間: 2025-3-25 17:43
Yongyi Shou,Mingu Kang,Young Won Parkcomputer and communication systems. While system . has received a lot of attention in the past, increasingly system . is gaining importance. The proliferation of computer and computer-based communication systems has contributed to this in no small measure. This chapter is thus a step in the directio作者: vasculitis 時(shí)間: 2025-3-25 22:43
https://doi.org/10.1007/978-1-4757-4828-4Markov Chains; Markov chain; Markov model; modeling; optimization作者: Instantaneous 時(shí)間: 2025-3-26 01:43
978-1-4419-5100-7Springer Science+Business Media New York 2000作者: absolve 時(shí)間: 2025-3-26 04:50
International Series in Operations Research & Management Sciencehttp://image.papertrans.cn/c/image/232907.jpg作者: 接合 時(shí)間: 2025-3-26 12:24
Computational Probability978-1-4757-4828-4Series ISSN 0884-8289 Series E-ISSN 2214-7934 作者: FILLY 時(shí)間: 2025-3-26 13:52
Robert Handfield,Daniel J. Finkenstadtges in the area of computational probability. To set the stage, we discuss the objectives of computational probability in more detail, and we point out the difficulties one encounters in this area. We also contrast computational probability with other approaches.作者: 歡笑 時(shí)間: 2025-3-26 17:33
Forschungsmodell Supply Chain Integration,is represented by a number of states and rules that govern the manner in which it moves from one state to the next. The state of an automaton at any time . is just the state it occupies at time . and the state of the SAN at time . is given by the state of each of its constituent automata.作者: conservative 時(shí)間: 2025-3-26 22:54
Helmuth Ludwig,Alastair Orchardchain to achieve an economic objective. Control is exercised by taking a sequence of actions, each of which may depend on the currently observed state and may influence both the immediate cost and the next state transition.作者: CT-angiography 時(shí)間: 2025-3-27 04:36 作者: annexation 時(shí)間: 2025-3-27 07:33
Stochastic Automata Networks,is represented by a number of states and rules that govern the manner in which it moves from one state to the next. The state of an automaton at any time . is just the state it occupies at time . and the state of the SAN at time . is given by the state of each of its constituent automata.作者: 和藹 時(shí)間: 2025-3-27 11:11
Optimal Control of Markov Chains,chain to achieve an economic objective. Control is exercised by taking a sequence of actions, each of which may depend on the currently observed state and may influence both the immediate cost and the next state transition.作者: 愛花花兒憤怒 時(shí)間: 2025-3-27 16:44 作者: Irascible 時(shí)間: 2025-3-27 19:36
Fallstudien Supply Chain Integration,of the M/G/1 queue, except that the entries are matrices. In the imbedded Markov chain of the GI/M/1 queue, all columns but the first have the same entries, except that they are displaced so that the diagonal block entry is common to all. Similarly, in the M/G/1 queue, all rows except the first one are equal after proper centering.作者: Scintigraphy 時(shí)間: 2025-3-27 22:37 作者: 講個(gè)故事逗他 時(shí)間: 2025-3-28 03:12 作者: 放肆的我 時(shí)間: 2025-3-28 07:23
Transient Solutions for Markov Chains,tions requires the calculation of measures during a relatively “short” period of time. These are the so-called .. In these cases the steady state measures are not good approximations for the transient, and one has to resort to different techniques to obtain the desired quantities.作者: BLAZE 時(shí)間: 2025-3-28 11:30 作者: Palpable 時(shí)間: 2025-3-28 17:10 作者: 打擊 時(shí)間: 2025-3-28 19:17
The Product form Tool for Queueing Networks,lled product form. This computationally attractive form will be shown to be directly related to the principle of balance per station. This principle will be used to provide practical insights concerning the following questions作者: 范例 時(shí)間: 2025-3-29 01:38
Book 2000systems, stochastic Petri-nets andsystems dealing with reliability - has benefited significantlyfrom these advances. The objective of this book is to make thesetopics accessible to researchers, graduate students, andpractitioners. Great care was taken to make the exposition as clear aspossible. Ever作者: Immobilize 時(shí)間: 2025-3-29 04:19 作者: PRISE 時(shí)間: 2025-3-29 09:49 作者: organism 時(shí)間: 2025-3-29 13:24
Techniques for System Dependability Evaluation,feration of computer and computer-based communication systems has contributed to this in no small measure. This chapter is thus a step in the direction of summarizing the techniques, tools and recent developments in the field of system dependability evaluation.作者: 河潭 時(shí)間: 2025-3-29 18:05 作者: Prophylaxis 時(shí)間: 2025-3-29 20:00
Use of Characteristic Roots for Solving Infinite State Markov Chains,ur focus in this chapter is on the discrete-time case, we will show how the continuous-time case can be handled by the same techniques. The M/G/1 and G/M/1 classes can be solved using the matrix analytic method [Neuts, 1981, Neuts, 1989], and we will also discuss the relationship between the approac作者: muscle-fibers 時(shí)間: 2025-3-30 01:35
An Introduction to Numerical Transform Inversion and Its Application to Probability Models,by Gaver [Gaver, 1966]. (For more on the history of numerical transform inversion, see our earlier survey [Abate and Whitt, 1992a].) Hence, in the application of probability models to engineering, transforms became regarded more as mathematical toys than practical tools. Indeed, the conventional wis作者: cipher 時(shí)間: 2025-3-30 04:41 作者: Militia 時(shí)間: 2025-3-30 08:38
Book 2000on matrices for Markov chains,with particular emphasis on stochastic Petri-nets. Chapter 3 discusseshow to find transient probabilities and transient rewards for theseMarkov chains. The next two chapters indicate how to find steady-stateprobabilities for Markov chains with a finite number of states.作者: 沉積物 時(shí)間: 2025-3-30 13:03
https://doi.org/10.1007/978-3-658-02664-6 is therefore quite simple. Unfortunately, problems arise from the computational point of view because of the large number of states which many systems may occupy. As indicated in Chapters 1 and 2, it is not uncommon for thousands of states to be generated even for simple applications.作者: 機(jī)制 時(shí)間: 2025-3-30 19:58
Helmuth Ludwig,Alastair Orchardur focus in this chapter is on the discrete-time case, we will show how the continuous-time case can be handled by the same techniques. The M/G/1 and G/M/1 classes can be solved using the matrix analytic method [Neuts, 1981, Neuts, 1989], and we will also discuss the relationship between the approac作者: CHART 時(shí)間: 2025-3-30 22:29