971 resultados para Logic outer-approximation algorithm
Resumo:
Teemanumero 1/2011 : Kauhut ja pelot.
Resumo:
The efficacy of three vaccines was evaluated in chickens for the control of experimental infection with Salmonella Enteritidis (SE) phage type 4. The vaccines were produced with bacterin, outer membrane proteins (OMP) and fimbriae crude extract (FE). The chickens were vaccinated intramuscularly with two doses of each vaccine at 12 and 15 weeks of age. The chickens were then orally challenged with 10(9) CFU/chicken Salmonella Enteritidis phage type 4 at 18 weeks of age. Fecal swabs were performed for the recovery of shedding SE, and SE was recovered from the liver and spleen. Additionally, antibody titers were measured in the serum by micro-agglutination test. The results indicated that the vaccine produced with bacterin yielded better results and resulted in reduction of fecal shedding and organ invasion by SE after oral challenge, although no vaccine was 100% effective for the control of SE experimental infection.
Resumo:
InsomniaGame oli Turun yliopiston digitaalisen kulttuurin oppiaineen ja Insomnia verkkopeliyhdistyksen yhteistyössä vuosina 2010 ja 2011 toteuttama pelikonseptikokeilu. InsomniaGame oli osa laajempaa ”CoEx: Yhteisöllistä tekemistä tukevat tilat kokemusten jakamisessa” kaksivuotista (1.10.2009–31.12.2011) hanketta, jonka toteuttivat yhteistyössä Turun yliopiston Porin yksikkö, Tampereen teknillisen yliopiston Porin yksikkö ja Tampereen yliopisto. Hankkeen tavoitteena oli toteuttaa sosiaalista mediaa, yhteisöllisyyttä ja lisättyä todellisuutta hyödyntäviä virtuaalisia ja julkisia tiloja, joissa käyttäjät voivat jakaa kokemuksia. Tutkimus on luonteeltaan soveltava pro gradu -tutkielma, joka sisältää kaksi vuotta kestäneen ja kaksi pelisovellusta sisältävän työosuuden. InsomniaGame koostui erilaisista pelaajien suorittamista tehtävistä, pelialustasta sekä taustatarinasta. Päätutkimuskysymykset ovat: Mitkä tekijät vaikuttivat pelisuunnitteluprosessiin ja miten? Työ esittelee InsomniaGame-pelin kehityksen. Erityistarkastelussa ovat suunnitteluprosessin ja pelin sisällölliset muutokset sekä niihin vaikuttaneet tekijät. Pelin kehitys perustui pääasiassa erilaisiin dokumentteihin, joita käytettiin suunnittelun apuvälineenä sekä viestinnässä projektin eri toimijoiden kesken. Tutkimus pyrkii syntyneiden dokumenttien sekä pelisuunnittelijoiden muistin perusteella rekonstruoimaan InsomniaGame-pelisovelluksen kehityskaaren. InsomniaGamen kehityksessä oli monia tekijöitä, jotka muuttuivat sen kehityskaaren aikana. Itse pelin sisältö, kuten myös suunnittelutapa, muuttuivat kahden vuoden aikana huomattavasti. Pelillä oli myös monia erityispiirteitä, jotka tekevät sen kehityksestä ainutlaatuisen, sillä esimerkiksi pelin testaaminen yhtenä kokonaisuutena oli mahdotonta. Lisäksi peli oli tutkimus- ja yhteistyöprojekti, jossa oli mukana monia eri toimijoita ja erityisesti tutkimuksessa korostuu yhteistyökumppani Insomnia verkkopeliyhdistyksen osallisuus. InsomniaGamen kummankaan vuoden toteutus ei sujunut odotetulla tavalla, mikä osaltaan vaikutti etenkin jälkimmäisen vuoden pelin suunnitteluun. Varsinainen suunnittelutyö kuitenkin eteni ensimmäisenä vuonna käytetyn mallin mukaisesti, mutta kuitenkin niin että alkuperäiset oletukset pelisuunnittelusta ja lopputuloksesta muuttuivat. Tämän vuoksi peliprojektia voi paikoitellen luonnehtia jopa kaoottiseksi, ja erityisesti toteutusvaiheessa jouduttiin luomaan nopealla aikataululla uusia toimintamalleja. Työ toimii mallina tuleville peliprojekteille, mutta erityisen tärkeää olisi luoda yhtenäinen kehitysalusta vastaavanlaisia projekteja varten.
Resumo:
This paper investigates defect detection methodologies for rolling element bearings through vibration analysis. Specifically, the utility of a new signal processing scheme combining the High Frequency Resonance Technique (HFRT) and Adaptive Line Enhancer (ALE) is investigated. The accelerometer is used to acquire data for this analysis, and experimental results have been obtained for outer race defects. Results show the potential effectiveness of the signal processing technique to determine both the severity and location of a defect. The HFRT utilizes the fact that much of the energy resulting from a defect impact manifests itself in the higher resonant frequencies of a system. Demodulation of these frequency bands through use of the envelope technique is then employed to gain further insight into the nature of the defect while further increasing the signal to noise ratio. If periodic, the defect frequency is then present in the spectra of the enveloped signal. The ALE is used to enhance the envelope spectrum by reducing the broadband noise. It provides an enhanced envelope spectrum with clear peaks at the harmonics of a characteristic defect frequency. It is implemented by using a delayed version of the signal and the signal itself to decorrelate the wideband noise. This noise is then rejected by the adaptive filter that is based upon the periodic information in the signal. Results have been obtained for outer race defects. They show the effectiveness of the methodology to determine both the severity and location of a defect. In two instances, a linear relationship between signal characteristics and defect size is indicated.
Resumo:
The determination of the intersection curve between Bézier Surfaces may be seen as the composition of two separated problems: determining initial points and tracing the intersection curve from these points. The Bézier Surface is represented by a parametric function (polynomial with two variables) that maps a point in the tridimensional space from the bidimensional parametric space. In this article, it is proposed an algorithm to determine the initial points of the intersection curve of Bézier Surfaces, based on the solution of polynomial systems with the Projected Polyhedral Method, followed by a method for tracing the intersection curves (Marching Method with differential equations). In order to allow the use of the Projected Polyhedral Method, the equations of the system must be represented in terms of the Bernstein basis, and towards this goal it is proposed a robust and reliable algorithm to exactly transform a multivariable polynomial in terms of power basis to a polynomial written in terms of Bernstein basis .
Resumo:
An Autonomous Mobile Robot battery driven, with two traction wheels and a steering wheel is being developed. This Robot central control is regulated by an IPC, which controls every function of security, steering, positioning localization and driving. Each traction wheel is operated by a DC motor with independent control system. This system is made up of a chopper, an encoder and a microcomputer. The IPC transmits the velocity values and acceleration ramp references to the PIC microcontrollers. As each traction wheel control is independent, it's possible to obtain different speed values for each wheel. This process facilities the direction and drive changes. Two different strategies for speed velocity control were implemented; one works with PID, and the other with fuzzy logic. There were no changes in circuits and feedback control, except for the PIC microcontroller software. Comparing the two different speed control strategies the results were equivalent. However, in relation to the development and implementation of these strategies, the difficulties were bigger to implement the PID control.
Resumo:
This work analyzes an active fuzzy logic control system in a Rijke type pulse combustor. During the system development, a study of the existing types of control for pulse combustion was carried out and a simulation model was implemented to be used with the package Matlab and Simulink. Blocks which were not available in the simulator library were developed. A fuzzy controller was developed and its membership functions and inference rules were established. The obtained simulation showed that fuzzy logic is viable in the control of combustion instabilities. The obtained results indicated that the control system responded to pulses in an efficient and desirable way. It was verified that the system needed approximately 0.2 s to increase the tube internal pressure from 30 to 90 mbar, with an assumed total delay of 2 ms. The effects of delay variation were studied. Convergence was always obtained and general performance was not affected by the delay. The controller sends a pressure signal in phase with the Rijke tube internal pressure signal, through the speakers, when an increase the oscillations pressure amplitude is desired. On the other hand, when a decrease of the tube internal pressure amplitude is desired, the controller sends a signal 180º out of phase.
Resumo:
This doctoral thesis introduces an improved control principle for active du/dt output filtering in variable-speed AC drives, together with performance comparisons with previous filtering methods. The effects of power semiconductor nonlinearities on the output filtering performance are investigated. The nonlinearities include the timing deviation and the voltage pulse waveform distortion in the variable-speed AC drive output bridge. Active du/dt output filtering (ADUDT) is a method to mitigate motor overvoltages in variable-speed AC drives with long motor cables. It is a quite recent addition to the du/dt reduction methods available. This thesis improves on the existing control method for the filter, and concentrates on the lowvoltage (below 1 kV AC) two-level voltage-source inverter implementation of the method. The ADUDT uses narrow voltage pulses having a duration in the order of a microsecond from an IGBT (insulated gate bipolar transistor) inverter to control the output voltage of a tuned LC filter circuit. The filter output voltage has thus increased slope transition times at the rising and falling edges, with an opportunity of no overshoot. The effect of the longer slope transition times is a reduction in the du/dt of the voltage fed to the motor cable. Lower du/dt values result in a reduction in the overvoltage effects on the motor terminals. Compared with traditional output filtering methods to accomplish this task, the active du/dt filtering provides lower inductance values and a smaller physical size of the filter itself. The filter circuit weight can also be reduced. However, the power semiconductor nonlinearities skew the filter control pulse pattern, resulting in control deviation. This deviation introduces unwanted overshoot and resonance in the filter. The controlmethod proposed in this thesis is able to directly compensate for the dead time-induced zero-current clamping (ZCC) effect in the pulse pattern. It gives more flexibility to the pattern structure, which could help in the timing deviation compensation design. Previous studies have shown that when a motor load current flows in the filter circuit and the inverter, the phase leg blanking times distort the voltage pulse sequence fed to the filter input. These blanking times are caused by excessively large dead time values between the IGBT control pulses. Moreover, the various switching timing distortions, present in realworld electronics when operating with a microsecond timescale, bring additional skew to the control. Left uncompensated, this results in distortion of the filter input voltage and a filter self-induced overvoltage in the form of an overshoot. This overshoot adds to the voltage appearing at the motor terminals, thus increasing the transient voltage amplitude at the motor. This doctoral thesis investigates the magnitude of such timing deviation effects. If the motor load current is left uncompensated in the control, the filter output voltage can overshoot up to double the input voltage amplitude. IGBT nonlinearities were observed to cause a smaller overshoot, in the order of 30%. This thesis introduces an improved ADUDT control method that is able to compensate for phase leg blanking times, giving flexibility to the pulse pattern structure and dead times. The control method is still sensitive to timing deviations, and their effect is investigated. A simple approach of using a fixed delay compensation value was tried in the test setup measurements. The ADUDT method with the new control algorithm was found to work in an actual motor drive application. Judging by the simulation results, with the delay compensation, the method should ultimately enable an output voltage performance and a du/dt reduction that are free from residual overshoot effects. The proposed control algorithm is not strictly required for successful ADUDT operation: It is possible to precalculate the pulse patterns by iteration and then for instance store them into a look-up table inside the control electronics. Rather, the newly developed control method is a mathematical tool for solving the ADUDT control pulses. It does not contain the timing deviation compensation (from the logic-level command to the phase leg output voltage), and as such is not able to remove the timing deviation effects that cause error and overshoot in the filter. When the timing deviation compensation has to be tuned-in in the control pattern, the precalculated iteration method could prove simpler and equally good (or even better) compared with the mathematical solution with a separate timing compensation module. One of the key findings in this thesis is the conclusion that the correctness of the pulse pattern structure, in the sense of ZCC and predicted pulse timings, cannot be separated from the timing deviations. The usefulness of the correctly calculated pattern is reduced by the voltage edge timing errors. The doctoral thesis provides an introductory background chapter on variable-speed AC drives and the problem of motor overvoltages and takes a look at traditional solutions for overvoltage mitigation. Previous results related to the active du/dt filtering are discussed. The basic operation principle and design of the filter have been studied previously. The effect of load current in the filter and the basic idea of compensation have been presented in the past. However, there was no direct way of including the dead time in the control (except for solving the pulse pattern manually by iteration), and the magnitude of nonlinearity effects had not been investigated. The enhanced control principle with the dead time handling capability and a case study of the test setup timing deviations are the main contributions of this doctoral thesis. The simulation and experimental setup results show that the proposed control method can be used in an actual drive. Loss measurements and a comparison of active du/dt output filtering with traditional output filtering methods are also presented in the work. Two different ADUDT filter designs are included, with ferrite core and air core inductors. Other filters included in the tests were a passive du/dtfilter and a passive sine filter. The loss measurements incorporated a silicon carbide diode-equipped IGBT module, and the results show lower losses with these new device technologies. The new control principle was measured in a 43 A load current motor drive system and was able to bring the filter output peak voltage from 980 V (the previous control principle) down to 680 V in a 540 V average DC link voltage variable-speed drive. A 200 m motor cable was used, and the filter losses for the active du/dt methods were 111W–126 W versus 184 W for the passive du/dt. In terms of inverter and filter losses, the active du/dt filtering method had a 1.82-fold increase in losses compared with an all-passive traditional du/dt output filter. The filter mass with the active du/dt method was 17% (2.4 kg, air-core inductors) compared with 14 kg of the passive du/dt method filter. Silicon carbide freewheeling diodes were found to reduce the inverter losses in the active du/dt filtering by 18% compared with the same IGBT module with silicon diodes. For a 200 m cable length, the average peak voltage at the motor terminals was 1050 V with no filter, 960 V for the all-passive du/dt filter, and 700 V for the active du/dt filtering applying the new control principle.
Resumo:
The dissertation proposes two control strategies, which include the trajectory planning and vibration suppression, for a kinematic redundant serial-parallel robot machine, with the aim of attaining the satisfactory machining performance. For a given prescribed trajectory of the robot's end-effector in the Cartesian space, a set of trajectories in the robot's joint space are generated based on the best stiffness performance of the robot along the prescribed trajectory. To construct the required system-wide analytical stiffness model for the serial-parallel robot machine, a variant of the virtual joint method (VJM) is proposed in the dissertation. The modified method is an evolution of Gosselin's lumped model that can account for the deformations of a flexible link in more directions. The effectiveness of this VJM variant is validated by comparing the computed stiffness results of a flexible link with the those of a matrix structural analysis (MSA) method. The comparison shows that the numerical results from both methods on an individual flexible beam are almost identical, which, in some sense, provides mutual validation. The most prominent advantage of the presented VJM variant compared with the MSA method is that it can be applied in a flexible structure system with complicated kinematics formed in terms of flexible serial links and joints. Moreover, by combining the VJM variant and the virtual work principle, a systemwide analytical stiffness model can be easily obtained for mechanisms with both serial kinematics and parallel kinematics. In the dissertation, a system-wide stiffness model of a kinematic redundant serial-parallel robot machine is constructed based on integration of the VJM variant and the virtual work principle. Numerical results of its stiffness performance are reported. For a kinematic redundant robot, to generate a set of feasible joints' trajectories for a prescribed trajectory of its end-effector, its system-wide stiffness performance is taken as the constraint in the joints trajectory planning in the dissertation. For a prescribed location of the end-effector, the robot permits an infinite number of inverse solutions, which consequently yields infinite kinds of stiffness performance. Therefore, a differential evolution (DE) algorithm in which the positions of redundant joints in the kinematics are taken as input variables was employed to search for the best stiffness performance of the robot. Numerical results of the generated joint trajectories are given for a kinematic redundant serial-parallel robot machine, IWR (Intersector Welding/Cutting Robot), when a particular trajectory of its end-effector has been prescribed. The numerical results show that the joint trajectories generated based on the stiffness optimization are feasible for realization in the control system since they are acceptably smooth. The results imply that the stiffness performance of the robot machine deviates smoothly with respect to the kinematic configuration in the adjacent domain of its best stiffness performance. To suppress the vibration of the robot machine due to varying cutting force during the machining process, this dissertation proposed a feedforward control strategy, which is constructed based on the derived inverse dynamics model of target system. The effectiveness of applying such a feedforward control in the vibration suppression has been validated in a parallel manipulator in the software environment. The experimental study of such a feedforward control has also been included in the dissertation. The difficulties of modelling the actual system due to the unknown components in its dynamics is noticed. As a solution, a back propagation (BP) neural network is proposed for identification of the unknown components of the dynamics model of the target system. To train such a BP neural network, a modified Levenberg-Marquardt algorithm that can utilize an experimental input-output data set of the entire dynamic system is introduced in the dissertation. Validation of the BP neural network and the modified Levenberg- Marquardt algorithm is done, respectively, by a sinusoidal output approximation, a second order system parameters estimation, and a friction model estimation of a parallel manipulator, which represent three different application aspects of this method.
Resumo:
Linguistic modelling is a rather new branch of mathematics that is still undergoing rapid development. It is closely related to fuzzy set theory and fuzzy logic, but knowledge and experience from other fields of mathematics, as well as other fields of science including linguistics and behavioral sciences, is also necessary to build appropriate mathematical models. This topic has received considerable attention as it provides tools for mathematical representation of the most common means of human communication - natural language. Adding a natural language level to mathematical models can provide an interface between the mathematical representation of the modelled system and the user of the model - one that is sufficiently easy to use and understand, but yet conveys all the information necessary to avoid misinterpretations. It is, however, not a trivial task and the link between the linguistic and computational level of such models has to be established and maintained properly during the whole modelling process. In this thesis, we focus on the relationship between the linguistic and the mathematical level of decision support models. We discuss several important issues concerning the mathematical representation of meaning of linguistic expressions, their transformation into the language of mathematics and the retranslation of mathematical outputs back into natural language. In the first part of the thesis, our view of the linguistic modelling for decision support is presented and the main guidelines for building linguistic models for real-life decision support that are the basis of our modeling methodology are outlined. From the theoretical point of view, the issues of representation of meaning of linguistic terms, computations with these representations and the retranslation process back into the linguistic level (linguistic approximation) are studied in this part of the thesis. We focus on the reasonability of operations with the meanings of linguistic terms, the correspondence of the linguistic and mathematical level of the models and on proper presentation of appropriate outputs. We also discuss several issues concerning the ethical aspects of decision support - particularly the loss of meaning due to the transformation of mathematical outputs into natural language and the issue or responsibility for the final decisions. In the second part several case studies of real-life problems are presented. These provide background and necessary context and motivation for the mathematical results and models presented in this part. A linguistic decision support model for disaster management is presented here – formulated as a fuzzy linear programming problem and a heuristic solution to it is proposed. Uncertainty of outputs, expert knowledge concerning disaster response practice and the necessity of obtaining outputs that are easy to interpret (and available in very short time) are reflected in the design of the model. Saaty’s analytic hierarchy process (AHP) is considered in two case studies - first in the context of the evaluation of works of art, where a weak consistency condition is introduced and an adaptation of AHP for large matrices of preference intensities is presented. The second AHP case-study deals with the fuzzified version of AHP and its use for evaluation purposes – particularly the integration of peer-review into the evaluation of R&D outputs is considered. In the context of HR management, we present a fuzzy rule based evaluation model (academic faculty evaluation is considered) constructed to provide outputs that do not require linguistic approximation and are easily transformed into graphical information. This is achieved by designing a specific form of fuzzy inference. Finally the last case study is from the area of humanities - psychological diagnostics is considered and a linguistic fuzzy model for the interpretation of outputs of multidimensional questionnaires is suggested. The issue of the quality of data in mathematical classification models is also studied here. A modification of the receiver operating characteristics (ROC) method is presented to reflect variable quality of data instances in the validation set during classifier performance assessment. Twelve publications on which the author participated are appended as a third part of this thesis. These summarize the mathematical results and provide a closer insight into the issues of the practicalapplications that are considered in the second part of the thesis.
Resumo:
This thesis is concerned with the state and parameter estimation in state space models. The estimation of states and parameters is an important task when mathematical modeling is applied to many different application areas such as the global positioning systems, target tracking, navigation, brain imaging, spread of infectious diseases, biological processes, telecommunications, audio signal processing, stochastic optimal control, machine learning, and physical systems. In Bayesian settings, the estimation of states or parameters amounts to computation of the posterior probability density function. Except for a very restricted number of models, it is impossible to compute this density function in a closed form. Hence, we need approximation methods. A state estimation problem involves estimating the states (latent variables) that are not directly observed in the output of the system. In this thesis, we use the Kalman filter, extended Kalman filter, Gauss–Hermite filters, and particle filters to estimate the states based on available measurements. Among these filters, particle filters are numerical methods for approximating the filtering distributions of non-linear non-Gaussian state space models via Monte Carlo. The performance of a particle filter heavily depends on the chosen importance distribution. For instance, inappropriate choice of the importance distribution can lead to the failure of convergence of the particle filter algorithm. In this thesis, we analyze the theoretical Lᵖ particle filter convergence with general importance distributions, where p ≥2 is an integer. A parameter estimation problem is considered with inferring the model parameters from measurements. For high-dimensional complex models, estimation of parameters can be done by Markov chain Monte Carlo (MCMC) methods. In its operation, the MCMC method requires the unnormalized posterior distribution of the parameters and a proposal distribution. In this thesis, we show how the posterior density function of the parameters of a state space model can be computed by filtering based methods, where the states are integrated out. This type of computation is then applied to estimate parameters of stochastic differential equations. Furthermore, we compute the partial derivatives of the log-posterior density function and use the hybrid Monte Carlo and scaled conjugate gradient methods to infer the parameters of stochastic differential equations. The computational efficiency of MCMC methods is highly depend on the chosen proposal distribution. A commonly used proposal distribution is Gaussian. In this kind of proposal, the covariance matrix must be well tuned. To tune it, adaptive MCMC methods can be used. In this thesis, we propose a new way of updating the covariance matrix using the variational Bayesian adaptive Kalman filter algorithm.
Resumo:
SDS, C12E8, CHAPS or CHAPSO or a combination of two of these detergents is generally used for the solubilization of Na,K-ATPase and other ATPases. Our method using only C12E8 has the advantage of considerable reduction of the time for enzyme purification, with rapid solubilization and purification in a single chromatographic step. Na,K-ATPase-rich membrane fragments of rabbit kidney outer medulla were obtained without adding SDS. Optimum conditions for solubilization were obtained at 4ºC after rapid mixing of 1 mg of membrane Na,K-ATPase with 1 mg of C12E8/ml, yielding 98% recovery of the activity. The solubilized enzyme was purified by gel filtration on a Sepharose 6B column at 4ºC. Non-denaturing PAGE revealed a single protein band with phosphomonohydrolase activity. The molecular mass of the purified enzyme estimated by gel filtration chromatography was 320 kDa. The optimum apparent pH obtained for the purified enzyme was 7.5 for both PNPP and ATP. The dependence of ATPase activity on ATP concentration showed high (K0.5 = 4.0 µM) and low (K0.5 = 1.4 mM) affinity sites for ATP, with negative cooperativity. Ouabain (5 mM), oligomycin (1 µg/ml) and sodium vanadate (3 µM) inhibited the ATPase activity of C12E8-solubilized and purified Na,K-ATPase by 99, 81 and 98.5%, respectively. We have shown that Na,K-ATPase solubilized only with C12E8 can be purified and retains its activity. The activity is consistent with the form of (alphaß)2 association.
Resumo:
This research work addresses the problem of building a mathematical model for the given system of heat exchangers and to determine the temperatures, pressures and velocities at the intermediate positions. Such model could be used in nding an optimal design for such a superstructure. To limit the size and computing time a reduced network model was used. The method can be generalized to larger network structures. A mathematical model which includes a system of non-linear equations has been built and solved according to the Newton-Raphson algorithm. The results obtained by the proposed mathematical model were compared with the results obtained by the Paterson approximation and Chen's Approximation. Results of this research work in collaboration with a current ongoing research at the department will optimize the valve positions and hence, minimize the pumping cost and maximize the heat transfer of the system of heat exchangers.
Resumo:
This thesis introduces an extension of Chomsky’s context-free grammars equipped with operators for referring to left and right contexts of strings.The new model is called grammar with contexts. The semantics of these grammars are given in two equivalent ways — by language equations and by logical deduction, where a grammar is understood as a logic for the recursive definition of syntax. The motivation for grammars with contexts comes from an extensive example that completely defines the syntax and static semantics of a simple typed programming language. Grammars with contexts maintain most important practical properties of context-free grammars, including a variant of the Chomsky normal form. For grammars with one-sided contexts (that is, either left or right), there is a cubic-time tabular parsing algorithm, applicable to an arbitrary grammar. The time complexity of this algorithm can be improved to quadratic,provided that the grammar is unambiguous, that is, it only allows one parsefor every string it defines. A tabular parsing algorithm for grammars withtwo-sided contexts has fourth power time complexity. For these grammarsthere is a recognition algorithm that uses a linear amount of space. For certain subclasses of grammars with contexts there are low-degree polynomial parsing algorithms. One of them is an extension of the classical recursive descent for context-free grammars; the version for grammars with contexts still works in linear time like its prototype. Another algorithm, with time complexity varying from linear to cubic depending on the particular grammar, adapts deterministic LR parsing to the new model. If all context operators in a grammar define regular languages, then such a grammar can be transformed to an equivalent grammar without context operators at all. This allows one to represent the syntax of languages in a more succinct way by utilizing context specifications. Linear grammars with contexts turned out to be non-trivial already over a one-letter alphabet. This fact leads to some undecidability results for this family of grammars