6 resultados para Distributed algorithm
em Archivo Digital para la Docencia y la Investigación - Repositorio Institucional de la Universidad del País Vasco
Resumo:
9 p.
Resumo:
Background: Primary distal renal tubular acidosis (dRTA) caused by mutations in the genes that codify for the H+ -ATPase pump subunits is a heterogeneous disease with a poor phenotype-genotype correlation. Up to now, large cohorts of dRTA Tunisian patients have not been analyzed, and molecular defects may differ from those described in other ethnicities. We aim to identify molecular defects present in the ATP6V1B1, ATP6V0A4 and SLC4A1 genes in a Tunisian cohort, according to the following algorithm: first, ATP6V1B1 gene analysis in dRTA patients with sensorineural hearing loss (SNHL) or unknown hearing status. Afterwards, ATP6V0A4 gene study in dRTA patients with normal hearing, and in those without any structural mutation in the ATP6V1B1 gene despite presenting SNHL. Finally, analysis of the SLC4A1 gene in those patients with a negative result for the previous studies. Methods: 25 children (19 boys) with dRTA from 20 families of Tunisian origin were studied. DNAs were extracted by the standard phenol/chloroform method. Molecular analysis was performed by PCR amplification and direct sequencing. Results: In the index cases, ATP6V1B1 gene screening resulted in a mutation detection rate of 81.25%, which increased up to 95% after ATP6V0A4 gene analysis. Three ATP6V1B1 mutations were observed: one frameshift mutation (c.1155dupC; p.Ile386fs), in exon 12; a G to C single nucleotide substitution, on the acceptor splicing site (c.175-1G > C; p.?) in intron 2, and one novel missense mutation (c. 1102G > A; p. Glu368Lys), in exon 11. We also report four mutations in the ATP6V0A4 gene: one single nucleotide deletion in exon 13 (c.1221delG; p. Met408Cysfs* 10); the nonsense c.16C > T; p.Arg6*, in exon 3; and the missense changes c.1739 T > C; p.Met580Thr, in exon 17 and c.2035G > T; p.Asp679Tyr, in exon 19. Conclusion: Molecular diagnosis of ATP6V1B1 and ATP6V0A4 genes was performed in a large Tunisian cohort with dRTA. We identified three different ATP6V1B1 and four different ATP6V0A4 mutations in 25 Tunisian children. One of them, c.1102G > A; p.Glu368Lys in the ATP6V1B1 gene, had not previously been described. Among deaf since childhood patients, 75% had the ATP6V1B1 gene c. 1155dupC mutation in homozygosis. Based on the results, we propose a new diagnostic strategy to facilitate the genetic testing in North Africans with dRTA and SNHL.
Resumo:
This doctoral Thesis defines and develops a new methodology for feeder reconfiguration in distribution networks with Distributed Energy Resources (DER). The proposed methodology is based on metaheuristic Ant Colony Optimization (ACO) algorithms. The methodology is called Item Oriented Ant System (IOAS) and the doctoral Thesis also defines three variations of the original methodology, Item Oriented Ant Colony System (IOACS), Item Oriented Max-min Ant System (IOMMAS) y Item Oriented Max-min Ant Colony System (IOACS). All methodologies pursue a twofold objective, to minimize the power losses and maximize DER penetration in distribution networks. The aim of the variations is to find the algorithm that adapts better to the present optimization problem, solving it most efficiently. The main feature of the methodology lies in the fact that the heuristic information and the exploitation information (pheromone) are attached to the item not to the path. Besides, the doctoral Thesis proposes to use feeder reconfiguration in order to increase the distribution network capacity of accepting a major degree of DER. The proposed methodology and its three variations have been tested and verified in two distribution networks well documented in the existing bibliography. These networks have been modeled and used to test all proposed methodologies for different scenarios with various DER penetration degrees.
Resumo:
This paper deals with the convergence of a remote iterative learning control system subject to data dropouts. The system is composed by a set of discrete-time multiple input-multiple output linear models, each one with its corresponding actuator device and its sensor. Each actuator applies the input signals vector to its corresponding model at the sampling instants and the sensor measures the output signals vector. The iterative learning law is processed in a controller located far away of the models so the control signals vector has to be transmitted from the controller to the actuators through transmission channels. Such a law uses the measurements of each model to generate the input vector to be applied to its subsequent model so the measurements of the models have to be transmitted from the sensors to the controller. All transmissions are subject to failures which are described as a binary sequence taking value 1 or 0. A compensation dropout technique is used to replace the lost data in the transmission processes. The convergence to zero of the errors between the output signals vector and a reference one is achieved as the number of models tends to infinity.
Resumo:
139 p.
Resumo:
Multi-Agent Reinforcement Learning (MARL) algorithms face two main difficulties: the curse of dimensionality, and environment non-stationarity due to the independent learning processes carried out by the agents concurrently. In this paper we formalize and prove the convergence of a Distributed Round Robin Q-learning (D-RR-QL) algorithm for cooperative systems. The computational complexity of this algorithm increases linearly with the number of agents. Moreover, it eliminates environment non sta tionarity by carrying a round-robin scheduling of the action selection and execution. That this learning scheme allows the implementation of Modular State-Action Vetoes (MSAV) in cooperative multi-agent systems, which speeds up learning convergence in over-constrained systems by vetoing state-action pairs which lead to undesired termination states (UTS) in the relevant state-action subspace. Each agent's local state-action value function learning is an independent process, including the MSAV policies. Coordination of locally optimal policies to obtain the global optimal joint policy is achieved by a greedy selection procedure using message passing. We show that D-RR-QL improves over state-of-the-art approaches, such as Distributed Q-Learning, Team Q-Learning and Coordinated Reinforcement Learning in a paradigmatic Linked Multi-Component Robotic System (L-MCRS) control problem: the hose transportation task. L-MCRS are over-constrained systems with many UTS induced by the interaction of the passive linking element and the active mobile robots.