首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Let be a positive integer, let F be a family of meromorphic functions in a domain D, all of whose zeros have multiplicity at least k+1, and let , be two holomorphic functions on D. If, for each fF, f=a(z)⇔f(k)=h(z), then F is normal in D.  相似文献   

2.
Normal families of meromorphic functions concerning shared values   总被引:2,自引:0,他引:2  
In this paper we study the problem of normal families of meromorphic functions concerning shared values and prove that a family F of meromorphic functions in a domain D is normal if for each pair of functions f and g in F, fafn and gagn share a value b in D where n is a positive integer and a,b are two finite constants such that n?4 and a≠0. This result is not true when n?3.  相似文献   

3.
We prove that if one or more players in a locally finite positional game have winning strategies, then they can find it by themselves, not losing more than a bounded number of plays and not using more than a linear-size memory, independently of the strategies applied by the other players. We design two algorithms for learning how to win. One of them can also be modified to determine a strategy that achieves a draw, provided that no winning strategy exists for the player in question but with properly chosen moves a draw can be ensured from the starting position. If a drawing- or winning strategy exists, then it is learnt after no more than a linear number of plays lost (linear in the number of edges of the game graph). Z. Tuza’s research has been supported in part by the grant OTKA T-049613.  相似文献   

4.
In this paper, we study the normality of a family of meromorphic functions and general criteria for normality of families of meromorphic functions with multiple zeros concerning shared values are obtained.  相似文献   

5.
The generalization of policies in reinforcement learning is a main issue, both from the theoretical model point of view and for their applicability. However, generalizing from a set of examples or searching for regularities is a problem which has already been intensively studied in machine learning. Thus, existing domains such as Inductive Logic Programming have already been linked with reinforcement learning. Our work uses techniques in which generalizations are constrained by a language bias, in order to regroup similar states. Such generalizations are principally based on the properties of concept lattices. To guide the possible groupings of similar states of the environment, we propose a general algebraic framework, considering the generalization of policies through a partition of the set of states and using a language bias as an a priori knowledge. We give a practical application as an example of our theoretical approach by proposing and experimenting a bottom-up algorithm.  相似文献   

6.
Industrial water systems often allow efficient water uses via water reuse and/or recirculation. The design of the network layout connecting water-using processes is a complex problem which involves several criteria to optimize. Most of the time, this design is achieved using Water Pinch technology, optimizing the freshwater flow rate entering the system. This paper describes an approach that considers two criteria: (i) the minimization of freshwater consumption and (ii) the minimization of the infrastructure cost required to build the network. The optimization model considers water reuse between operations and wastewater treatment as the main mechanisms to reduce freshwater consumption. The model is solved using multi-objective distributed Q-learning (MDQL), a heuristic approach based on the exploitation of knowledge acquired during the search process. MDQL has been previously tested on several multi-objective optimization benchmark problems with promising results [C. Mariano, Reinforcement learning in multi-objective optimization, Ph.D. thesis in Computer Science, Instituto Tecnológico y de Estudios Superiores de Monterrey, Campus Cuernavaca, March, 2002, Cuernavaca, Mor., México, 2001]. In order to compare the quality of the results obtained with MDQL, the reduced gradient method was applied to solve a weighted combination of the two objective functions used in the model. The proposed approach was tested on three cases: (i) a single contaminant four unitary operations problem where freshwater consumption is reduced via water reuse, (ii) a four contaminants real-world case with ten unitary operations, also with water reuse, and (iii) the water distribution network operation of Cuernavaca, Mexico, considering reduction of water leaks, operation of existing treatment plants at their design capacity, and design and construction of new treatment infrastructure to treat 100% of the wastewater produced. It is shown that the proposed approach can solved highly constrained real-world multi-objective optimization problems.  相似文献   

7.
In this paper, we investigate the reliability of a type of 1-for-2 shared protection systems. The 1-for-2 shared protection system is the most basic fault-tolerant configuration with shared backup units. We assume that there are two working units each serving a single user and one shared protection (spare) unit in the system. We also assume that the times to failure and to repair are subject to exponential and general distributions respectively. Under these assumptions, we derive the Laplace transform of the survival function (the cdf that the system will survive beyond a given time) for each user as well as the user-perceived Mean Time to First Failure (MTTFF) by combining the state transition analysis and the supplementary variable method. We also show the effect of the repair-time distribution, the failure rates and the repair rates of the units through the case study of small-sized two enterprises that share one spare device for backup purpose. The analysis reveals what is important and what should be done in order to improve the user-perceived reliability of shared protection systems.  相似文献   

8.
We consider a single-product make-to-stock manufacturing–remanufacturing system. Returned products require remanufacturing before they can be sold. The manufacturing and remanufacturing operations are executed by the same single server, where switching from one activity to another does not involve time or cost and can be done at an arbitrary moment in time. Customer demand can be fulfilled by either newly manufactured or remanufactured products. The times for manufacturing and remanufacturing a product are exponentially distributed. Demand and used products arrive via mutually independent Poisson processes. Disposal of products is not allowed and all used products that are returned have to be accepted. Using Markov decision processes, we investigate the optimal manufacture–remanufacture policy that minimizes holding, backorder, manufacturing and remanufacturing costs per unit of time over an infinite horizon. For a subset of system parameter values we are able to completely characterize the optimal continuous-review dynamic preemptive policy. We provide an efficient algorithm based on quasi-birth–death processes to compute the optimal policy parameter values. For other sets of system parameter values, we present some structural properties and insights related to the optimal policy and the performance of some simple threshold policies.  相似文献   

9.
In this paper, we use reinforcement learning (RL) techniques to determine dynamic prices in an electronic monopolistic retail market. The market that we consider consists of two natural segments of customers, captives and shoppers. Captives are mature, loyal buyers whereas the shoppers are more price sensitive and are attracted by sales promotions and volume discounts. The seller is the learning agent in the system and uses RL to learn from the environment. Under (reasonable) assumptions about the arrival process of customers, inventory replenishment policy, and replenishment lead time distribution, the system becomes a Markov decision process thus enabling the use of a wide spectrum of learning algorithms. In this paper, we use the Q-learning algorithm for RL to arrive at optimal dynamic prices that optimize the seller’s performance metric (either long term discounted profit or long run average profit per unit time). Our model and methodology can also be used to compute optimal reorder quantity and optimal reorder point for the inventory policy followed by the seller and to compute the optimal volume discounts to be offered to the shoppers.  相似文献   

10.
Learning Classifier Systems (LCS) are rule based Reinforcement Learning (RL) systems which use a generalization capability. In this paper, we highlight the differences between two kinds of LCSs. Some are used to directly perform RL while others latently learn a model of the interactions between the agent and its environment. Such a model can be used to speed up the core RL process. Thus, these two kinds of learning processes are complementary. We show here how the notion of generalization differs depending on whether the system anticipates (like Anticipatory Classifier System (ACS) and Yet Another Classifier System (YACS)) or not (like XCS). Moreover, we show some limitations of the formalism common to ACS and YACS, and propose a new system, called Modular Anticipatory Classifier System (MACS), which allows the latent learning process to take advantage of new regularities. We describe how the model can be used to perform active exploration and how this exploration may be aggregated with the policy resulting from the reinforcement learning process. The different algorithms are validated experimentally and some limitations in presence of uncertainties are highlighted.  相似文献   

11.
A model is designed and used to simulate how partners in a supplyrelationship identify and reach a common target in the form of an ideal endproduct. They cooperate fully and share returns. They learn by interaction,as follows. From their different perspectives, they complement each other'sidentification of the target. They adapt their productive competencies tothe target, in order to conform to demand (quality), and to each other, inorder to achieve efficient complementarity in production (efficiency). Asthey approach the target, their accuracy of identifying the targetincreases. Also, their speed of adaptation increases, and thus they can besaid to be learning by doing. The model allows two different patterns ofacceleration: a routine and a radical type of development. At some distancefrom the target they start to produce. A longer distance from the targetyields earlier returns, but also entails a greater compromise on quality andthereby yields lower returns. Unpredictable changes in market and technologyyield random shifts of the target. In the analysis, the returns from singleand dual sourcing are compared under different parameter settings. Thesimulations show that in line with expectations dual sourcing can be moreadvantageous if development is of the radical type. However, the advantageonly arises if conditions of market and technology are neither too volatilenor too stable.  相似文献   

12.
Accelerating autonomous learning by using heuristic selection of actions   总被引:2,自引:0,他引:2  
This paper investigates how to make improved action selection for online policy learning in robotic scenarios using reinforcement learning (RL) algorithms. Since finding control policies using any RL algorithm can be very time consuming, we propose to combine RL algorithms with heuristic functions for selecting promising actions during the learning process. With this aim, we investigate the use of heuristics for increasing the rate of convergence of RL algorithms and contribute with a new learning algorithm, Heuristically Accelerated Q-learning (HAQL), which incorporates heuristics for action selection to the Q-Learning algorithm. Experimental results on robot navigation show that the use of even very simple heuristic functions results in significant performance enhancement of the learning rate.  相似文献   

13.
分析极限理论在数学分析学习中的重要性,介绍学习极限理论的三点经验与体会,即充分理解基本内容;加强基本功训练;试着多问自己一些问题,学会举出各种各样的例子。  相似文献   

14.
论文介绍了计划评审法(PERT)的基本思想,探讨了学习曲线模型在PERT图中的应用,并给出了相应实例.  相似文献   

15.
We study an agency model, in which the principal has only incomplete information about the agent's preferences, in a dynamic setting. Through repeated interaction with the agent, the principal learns about the agent's preferences and can thus adjust the inventive system. In a dynamic computational model, we compare different learning strategies of the principal when facing different types of agents. The results indicate that better learning of preferences can improve the situation of both parties, but the learning process is rather sensitive to random disturbances.  相似文献   

16.
This paper presents a dynamical theory of organizational learning that explicitly accounts for both the introduction of new routines into the manufacturing process and improvements in the selection of which procedures to follow. Besides producing a power law of organizational learning with a rate that depends on the effectiveness of the decision procedure, the theory also accounts for observed anomalies characterized by price increases with cumulative output.  相似文献   

17.
研究性学习的关键问题之一是充分调动学生主动学习的兴趣与热情,提高学生学习效率,变被动学习为主动学习.结合实例给出了四种激发学生学习兴趣的研究性学习方法,即逆向思维、发散思维、基于Matlab的验证学习以及基于实际问题的教学研究方法.这四种学习方式的有机结合能在理论和实践上有效激发学生主动学习热情及学习兴趣,从而大大提高学生学习效率.  相似文献   

18.
针对基于机器学习的传统验证码识别受字符分割限制与人工操作过多等问题,基于深度学习Tensorflow学习框架将卷积神经网络应用到验证码的特性提取、分析、归类和识别中.将图片验证码作为整体输入,改进传统的LeNet-5网络结构,构建一种端到端的9层卷积神经网络,对验证码图像由低级到高级逐层提取图像特征,实现对图片验证码的识别.模型确定后采用控制变量法,针对每一迭代次数所处理的图片数量进行分析,对其准确率、损失值、训练时间进行综合测评,最终选取最优参数.实验结果显示,每批次处理128张图片,每迭代次数用时6秒,准确率的上限最高达到92%,损失值的下限最低达到0.0184.  相似文献   

19.
This article proposes a penalized likelihood method to jointly estimate multiple precision matrices for use in quadratic discriminant analysis (QDA) and model-based clustering. We use a ridge penalty and a ridge fusion penalty to introduce shrinkage and promote similarity between precision matrix estimates. We use blockwise coordinate descent for optimization, and validation likelihood is used for tuning parameter selection. Our method is applied in QDA and semi-supervised model-based clustering.  相似文献   

20.
Attention deficit hyperactivity disorder (ADHD) is characterized by decreased attention span, impulsiveness, and hyperactivity. Autonomic nervous system imbalance was previously described in this population. We aim to compare the autonomic function of children with ADHD and controls by analyzing heart rate variability (HRV). Children with ADHD (22 boys, mean age 9.964 years) and 28 controls (15 boys, mean age 9.857 years) rested in supine position with spontaneous breathing for 20 min. Heart rate was recorded beat by beat. HRV analysis was performed by use of chaotic global techniques. ADHD promoted an increase in the chaotic forward parameter. The algorithm which applied all three chaotic global parameters was only the second optimum statistically measured by Kruskal–Wallis (P < 0.0001) and low standard deviations. It was also highly influential by principal component analysis with almost all variation covered by the first two components. The third algorithm which lacked the (high spectral Detrended Fluctuation Analysis) parameter performed best statistically. However, we chose the algorithm which applied all three chaotic globals due to previous studies mentioned in the text—forward and inverse problems. Comparison of the autonomic function by analyzing HRV with chaotic global techniques suggests an increase in chaotic activity in children with ADHD in relation to the control group. © 2015 Wiley Periodicals, Inc. Complexity 21: 412–419, 2016  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号