835 resultados para Modified FDD Design Technology


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Cu(II)/neocuproine (2,9-dimethyl-1,10-phenanthroline) complexes were utilized for spectrophotometric determination of total polyphenol content in beers. This procedure is based on the reduction of Cu(II) by polyphenols in hydroethanolic medium (pH 7.0) in the presence of neocuproine, yielding Cu(I)/complexes with maximum absorption at 454 nm. The sensitivity of the proposed method was compared with the AOAC method using tannic and gallic acid as standards. The average apparent molar absorptivity, in L cm(-1) mol(-1), of tannic acid (3.50 +/- 0.20) x 10(5) and gallic acid (5.12 +/- 0.21) x 10(4), was twice as high for the proposed method. A lower limit of detection (LOD) (2.9 x 10(-1) mg L-1) was found when tannic acid was used in the proposed method. Additionally, less interference from the most common additives in beers was noticed. Total content of polyphenols was analyzed in 17 Brazilian samples. Results ranged from 35.5 to 556 mg L-1 of tannic acid, with higher values for recovery rates (45.4-118%, mean 85.0%) than for gallic acid. Although tannic acid is a mixture of polygalloyl glucoses, the total polyphenol content found in the samples suggests that tannic acid should be used as a standard. This is the first attempt to use this particular copper complex to quantify total polyphenol content in beer samples. (C) 2012 Elsevier Inc. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The microencapsulation of Lippia sidoides extracts in blends of carbohydrates was investigated. The extraction conditions were determined through a 2(2) factorial design. The effects of the plant:solvent ratio (A - 7.5:100 and 15:100 m/m) and the extraction time (B - 30 and 90 min) on thymol content of extractive solutions were evaluated, using a 2:1 (v/v) of ethanol:water at a temperature of 50 degrees C, as a solvent system. The selected extract was subjected to spray drying. Blends of maltodextrin and gum arabic at different proportions (4:1; 3:2; 2:3; 0:1) (m/m) were used as encapsulating material. The protective effects of the maltodextrin and gum arabic blends were evaluated by determination of the thymol retention in the dried product, which ranged from 70.2 to 84.2% (related to the content in the extractive solution). An increase in the gum arabic to maltodextrin (DE10) ratio has positive effect on thymol retention. L. sidoides extracts and spray-dried products showed antifungal activity against tested fungal strains (Candida albicans - ATCC 64548, Candida glabrata - ATCC 90030, Candida krusei - ATCC 6258, and Candida parapsilosis - ATCC 22019), evidencing their potential as a natural antifungal agent for medicinal, food, and cosmeceutical purposes. (C) 2012 The Institution of Chemical Engineers. Published by Elsevier B.V. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Microencapsulation can be an alternative to minimize lycopene instability. Thus, the aim of this study was to microencapsulate lycopene by spray drying, using a modified starch (Capsul (R)) as an encapsulating agent, and to assess the functionality of the capsules applying them in cake. The quantity of lycopene was varied at 5, 10 and 15% in a solution containing 30% of solids in order to obtain the microcapsules. These microcapsules were evaluated as to encapsulation efficiency and morphology and then submitted to a stability test and applied in cakes. Encapsulation efficiency values varied between 21 and 29%. The microcapsules had a rounded outer surface with the formation of concavities and they varied in size. The stability test revealed that microencapsulation offered greater protection to lycopene compared to its free form and it was observed that the microcapsules were able to release pigment and color the studied food system in a homogenous manner. (C) 2011 The Institution of Chemical Engineers. Published by Elsevier B.V. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the field of vehicle dynamics, commercial software can aid the designer during the conceptual and detailed design phases. Simulations using these tools can quickly provide specific design metrics, such as yaw and lateral velocity, for standard maneuvers. However, it remains challenging to correlate these metrics with empirical quantities that depend on many external parameters and design specifications. This scenario is the case with tire wear, which depends on the frictional work developed by the tire-road contact. In this study, an approach is proposed to estimate the tire-road friction during steady-state longitudinal and cornering maneuvers. Using this approach, a qualitative formula for tire wear evaluation is developed, and conceptual design analyses of cornering maneuvers are performed using simplified vehicle models. The influence of some design parameters such as cornering stiffness, the distance between the axles, and the steer angle ratio between the steering axles for vehicles with two steering axles is evaluated. The proposed methodology allows the designer to predict tire wear using simplified vehicle models during the conceptual design phase.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This work aimed at evaluating the spray congealing method for the production of microparticles of carbamazepine combined with a polyoxylglyceride carrier. In addition, the influence of the spray congealing conditions on the improvement of drug solubility was investigated using a three-factor, three-level Box-Behnken design. The factors studied were the cooling air flow rate, atomizing pressure, and molten dispersion feed rate. Dependent variables were the yield, solubility, encapsulation efficiency, particle size, water activity, and flow properties. Statistical analysis showed that only the yield was affected by the factors studied. The characteristics of the microparticles were evaluated using X-ray powder diffraction, scanning electron microscopy, differential scanning calorimetry, and hot-stage microscopy. The results showed a spherical morphology and changes in the crystalline state of the drug. The microparticles were obtained with good yields and encapsulation efficiencies, which ranged from 50 to 80% and 99.5 to 112%, respectively. The average size of the microparticles ranged from 17.7 to 39.4 mu m, the water activities were always below 0.5, and flowability was good to moderate. Both the solubility and dissolution rate of carbamazepine from the spray congealed microparticles were remarkably improved. The carbamazepine solubility showed a threefold increase and dissolution profile showed a twofold increase after 60 min compared to the raw drug. The Box-Behnken fractional factorial design proved to be a powerful tool to identify the best conditions for the manufacture of solid dispersion microparticles by spray congealing.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a technique for performing analog design synthesis at circuit level providing feedback to the designer through the exploration of the Pareto frontier. A modified simulated annealing which is able to perform crossover with past anchor points when a local minimum is found which is used as the optimization algorithm on the initial synthesis procedure. After all specifications are met, the algorithm searches for the extreme points of the Pareto frontier in order to obtain a non-exhaustive exploration of the Pareto front. Finally, multi-objective particle swarm optimization is used to spread the results and to find a more accurate frontier. Piecewise linear functions are used as single-objective cost functions to produce a smooth and equal convergence of all measurements to the desired specifications during the composition of the aggregate objective function. To verify the presented technique two circuits were designed, which are: a Miller amplifier with 96 dB Voltage gain, 15.48 MHz unity gain frequency, slew rate of 19.2 V/mu s with a current supply of 385.15 mu A, and a complementary folded cascode with 104.25 dB Voltage gain, 18.15 MHz of unity gain frequency and a slew rate of 13.370 MV/mu s. These circuits were synthesized using a 0.35 mu m technology. The results show that the method provides a fast approach for good solutions using the modified SA and further good Pareto front exploration through its connection to the particle swarm optimization algorithm.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Presented herein is the design of a dinuclear Ni-II synthetic hydrolase [Ni-2(HBPPAMFF)(mu-OAc)(2)(H2O)]-BPh4 (1) (H(2)BPPAMFF = 2-[(N-benzyl-N-2-pyridylmethylamine)]-4-methyl-6-[N-(2-pyridylmethyl)aminomethyl)])-4- methyl-6-formylphenol) to be covalently attached to silica surfaces, while maintaining its catalytic activity. An aldehyde-containing ligand (H(2)BPPAMFF) provides a reactive functional group that can serve as a cross-linking group to bind the complex to an organoalkoxysilane and later to the silica surfaces or directly to amino-modified surfaces. The dinuclear Ni-II complex covalently attached to the silica surfaces was fully characterized by different techniques. The catalytic turnover number (k(cat)) of the immobilized (NiNiII)-Ni-II catalyst in the hydrolysis of 2,4-bis(dinitrophenyl)phosphate is comparable to the homogeneous reaction; however, the catalyst interaction with the support enhanced the substrate to complex association constant, and consequently, the catalytic efficiency (E - k(cat)/K-M) and the supported catalyst can be reused for subsequent diester hydrolysis reactions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Cutting and packing problems are found in numerous industries such as garment, wood and shipbuilding. The collision free region concept is presented, as it represents all the translations possible for an item to be inserted into a container with already placed items. The often adopted nofit polygon concept and its analogous concept inner fit polygon are used to determine the collision free region. Boolean operations involving nofit polygons and inner fit polygons are used to determine the collision free region. New robust non-regularized Boolean operations algorithm is proposed to determine the collision free region. The algorithm is capable of dealing with degenerated boundaries. This capability is important because degenerated boundaries often represent local optimal placements. A parallelized version of the algorithm is also proposed and tests are performed in order to determine the execution times of both the serial and parallel versions of the algorithm.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Abstract This paper describes a design methodology for piezoelectric energy harvester s that thinly encapsulate the mechanical devices and expl oit resonances from higher- order vibrational modes. The direction of polarization determines the sign of the pi ezoelectric tensor to avoid cancellations of electric fields from opposite polarizations in the same circuit. The resultant modified equations of state are solved by finite element method (FEM). Com- bining this method with the solid isotropic material with penalization (SIMP) method for piezoelectric material, we have developed an optimization methodology that optimizes the piezoelectric material layout and polarization direc- tion. Updating the density function of the SIMP method is performed based on sensitivity analysis, the sequen- tial linear programming on the early stage of the opti- mization, and the phase field method on the latter stage

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The world of communication has changed quickly in the last decade resulting in the the rapid increase in the pace of peoples’ lives. This is due to the explosion of mobile communication and the internet which has now reached all levels of society. With such pressure for access to communication there is increased demand for bandwidth. Photonic technology is the right solution for high speed networks that have to supply wide bandwidth to new communication service providers. In particular this Ph.D. dissertation deals with DWDM optical packet-switched networks. The issue introduces a huge quantity of problems from physical layer up to transport layer. Here this subject is tackled from the network level perspective. The long term solution represented by optical packet switching has been fully explored in this years together with the Network Research Group at the department of Electronics, Computer Science and System of the University of Bologna. Some national as well as international projects supported this research like the Network of Excellence (NoE) e-Photon/ONe, funded by the European Commission in the Sixth Framework Programme and INTREPIDO project (End-to-end Traffic Engineering and Protection for IP over DWDM Optical Networks) funded by the Italian Ministry of Education, University and Scientific Research. Optical packet switching for DWDM networks is studied at single node level as well as at network level. In particular the techniques discussed are thought to be implemented for a long-haul transport network that connects local and metropolitan networks around the world. The main issues faced are contention resolution in a asynchronous variable packet length environment, adaptive routing, wavelength conversion and node architecture. Characteristics that a network must assure as quality of service and resilience are also explored at both node and network level. Results are mainly evaluated via simulation and through analysis.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Introduction 1.1 Occurrence of polycyclic aromatic hydrocarbons (PAH) in the environment Worldwide industrial and agricultural developments have released a large number of natural and synthetic hazardous compounds into the environment due to careless waste disposal, illegal waste dumping and accidental spills. As a result, there are numerous sites in the world that require cleanup of soils and groundwater. Polycyclic aromatic hydrocarbons (PAHs) are one of the major groups of these contaminants (Da Silva et al., 2003). PAHs constitute a diverse class of organic compounds consisting of two or more aromatic rings with various structural configurations (Prabhu and Phale, 2003). Being a derivative of benzene, PAHs are thermodynamically stable. In addition, these chemicals tend to adhere to particle surfaces, such as soils, because of their low water solubility and strong hydrophobicity, and this results in greater persistence under natural conditions. This persistence coupled with their potential carcinogenicity makes PAHs problematic environmental contaminants (Cerniglia, 1992; Sutherland, 1992). PAHs are widely found in high concentrations at many industrial sites, particularly those associated with petroleum, gas production and wood preserving industries (Wilson and Jones, 1993). 1.2 Remediation technologies Conventional techniques used for the remediation of soil polluted with organic contaminants include excavation of the contaminated soil and disposal to a landfill or capping - containment - of the contaminated areas of a site. These methods have some drawbacks. The first method simply moves the contamination elsewhere and may create significant risks in the excavation, handling and transport of hazardous material. Additionally, it is very difficult and increasingly expensive to find new landfill sites for the final disposal of the material. The cap and containment method is only an interim solution since the contamination remains on site, requiring monitoring and maintenance of the isolation barriers long into the future, with all the associated costs and potential liability. A better approach than these traditional methods is to completely destroy the pollutants, if possible, or transform them into harmless substances. Some technologies that have been used are high-temperature incineration and various types of chemical decomposition (for example, base-catalyzed dechlorination, UV oxidation). However, these methods have significant disadvantages, principally their technological complexity, high cost , and the lack of public acceptance. Bioremediation, on the contrast, is a promising option for the complete removal and destruction of contaminants. 1.3 Bioremediation of PAH contaminated soil & groundwater Bioremediation is the use of living organisms, primarily microorganisms, to degrade or detoxify hazardous wastes into harmless substances such as carbon dioxide, water and cell biomass Most PAHs are biodegradable unter natural conditions (Da Silva et al., 2003; Meysami and Baheri, 2003) and bioremediation for cleanup of PAH wastes has been extensively studied at both laboratory and commercial levels- It has been implemented at a number of contaminated sites, including the cleanup of the Exxon Valdez oil spill in Prince William Sound, Alaska in 1989, the Mega Borg spill off the Texas coast in 1990 and the Burgan Oil Field, Kuwait in 1994 (Purwaningsih, 2002). Different strategies for PAH bioremediation, such as in situ , ex situ or on site bioremediation were developed in recent years. In situ bioremediation is a technique that is applied to soil and groundwater at the site without removing the contaminated soil or groundwater, based on the provision of optimum conditions for microbiological contaminant breakdown.. Ex situ bioremediation of PAHs, on the other hand, is a technique applied to soil and groundwater which has been removed from the site via excavation (soil) or pumping (water). Hazardous contaminants are converted in controlled bioreactors into harmless compounds in an efficient manner. 1.4 Bioavailability of PAH in the subsurface Frequently, PAH contamination in the environment is occurs as contaminants that are sorbed onto soilparticles rather than in phase (NAPL, non aqueous phase liquids). It is known that the biodegradation rate of most PAHs sorbed onto soil is far lower than rates measured in solution cultures of microorganisms with pure solid pollutants (Alexander and Scow, 1989; Hamaker, 1972). It is generally believed that only that fraction of PAHs dissolved in the solution can be metabolized by microorganisms in soil. The amount of contaminant that can be readily taken up and degraded by microorganisms is defined as bioavailability (Bosma et al., 1997; Maier, 2000). Two phenomena have been suggested to cause the low bioavailability of PAHs in soil (Danielsson, 2000). The first one is strong adsorption of the contaminants to the soil constituents which then leads to very slow release rates of contaminants to the aqueous phase. Sorption is often well correlated with soil organic matter content (Means, 1980) and significantly reduces biodegradation (Manilal and Alexander, 1991). The second phenomenon is slow mass transfer of pollutants, such as pore diffusion in the soil aggregates or diffusion in the organic matter in the soil. The complex set of these physical, chemical and biological processes is schematically illustrated in Figure 1. As shown in Figure 1, biodegradation processes are taking place in the soil solution while diffusion processes occur in the narrow pores in and between soil aggregates (Danielsson, 2000). Seemingly contradictory studies can be found in the literature that indicate the rate and final extent of metabolism may be either lower or higher for sorbed PAHs by soil than those for pure PAHs (Van Loosdrecht et al., 1990). These contrasting results demonstrate that the bioavailability of organic contaminants sorbed onto soil is far from being well understood. Besides bioavailability, there are several other factors influencing the rate and extent of biodegradation of PAHs in soil including microbial population characteristics, physical and chemical properties of PAHs and environmental factors (temperature, moisture, pH, degree of contamination). Figure 1: Schematic diagram showing possible rate-limiting processes during bioremediation of hydrophobic organic contaminants in a contaminated soil-water system (not to scale) (Danielsson, 2000). 1.5 Increasing the bioavailability of PAH in soil Attempts to improve the biodegradation of PAHs in soil by increasing their bioavailability include the use of surfactants , solvents or solubility enhancers.. However, introduction of synthetic surfactant may result in the addition of one more pollutant. (Wang and Brusseau, 1993).A study conducted by Mulder et al. showed that the introduction of hydropropyl-ß-cyclodextrin (HPCD), a well-known PAH solubility enhancer, significantly increased the solubilization of PAHs although it did not improve the biodegradation rate of PAHs (Mulder et al., 1998), indicating that further research is required in order to develop a feasible and efficient remediation method. Enhancing the extent of PAHs mass transfer from the soil phase to the liquid might prove an efficient and environmentally low-risk alternative way of addressing the problem of slow PAH biodegradation in soil.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recent developments in piston engine technology have increased performance in a very significant way. Diesel turbocharged/turbo compound engines, fuelled by jet fuels, have great performances. The focal point of this thesis is the transformation of the FIAT 1900 jtd diesel common rail engine for the installation on general aviation aircrafts like the CESSNA 172. All considerations about the diesel engine are supported by the studies that have taken place in the laboratories of the II Faculty of Engineering in Forlì. This work, mostly experimental, concerns the transformation of the automotive FIAT 1900 jtd – 4 cylinders – turbocharged – diesel common rail into an aircraft engine. The design philosophy of the aluminium alloy basement of the spark ignition engine have been transferred to the diesel version while the pistons and the head of the FIAT 1900 jtd are kept in the aircraft engine. Different solutions have been examined in this work. A first V 90° cylinders version that can develop up to 300 CV and whose weight is 30 kg, without auxiliaries and turbocharging group. The second version is a development of e original version of the diesel 1900 cc engine with an optimized crankshaft, that employ a special steel, 300M, and that is verified for the aircraft requirements. Another version with an augmented stroke and with a total displacement of 2500 cc has been examined; the result is a 30% engine heavier. The last version proposed is a 1600 cc diesel engine that work at 5000 rpm, with a reduced stroke and capable of more than 200 CV; it was inspired to the Yamaha R1 motorcycle engine. The diesel aircraft engine design keeps the bore of 82 mm, while the stroke is reduced to 64.6 mm, so the engine size is reduced along with weight. The basement weight, in GD AlSi 9 MgMn alloy, is 8,5 kg. Crankshaft, rods and accessories have been redesigned to comply to aircraft standards. The result is that the overall size is increased of only the 8% when referred to the Yamaha engine spark ignition version, while the basement weight increases of 53 %, even if the bore of the diesel version is 11% lager. The original FIAT 1900 jtd piston has been slightly modified with the combustion chamber reworked to the compression ratio of 15:1. The material adopted for the piston is the aluminium alloy A390.0-T5 commonly used in the automotive field. The piston weight is 0,5 kg for the diesel engine. The crankshaft is verified to torsional vibrations according to the Lloyd register of shipping requirements. The 300M special steel crankshaft total weight is of 14,5 kg. The result reached is a very small and light engine that may be certified for general aviation: the engine weight, without the supercharger, air inlet assembly, auxiliary generators and high pressure body, is 44,7 kg and the total engine weight, with enlightened HP pump body and the titanium alloy turbocharger is less than 100 kg, the total displacement is 1365 cm3 and the estimated output power is 220 CV. The direct conversion of automotive piston engine to aircrafts pays too huge weight penalties. In fact the main aircraft requirement is to optimize the power to weight ratio in order to obtain compact and fast engines for aeronautical use: this 1600 common rail diesel engine version demonstrates that these results can be reached.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the past decade, the advent of efficient genome sequencing tools and high-throughput experimental biotechnology has lead to enormous progress in the life science. Among the most important innovations is the microarray tecnology. It allows to quantify the expression for thousands of genes simultaneously by measurin the hybridization from a tissue of interest to probes on a small glass or plastic slide. The characteristics of these data include a fair amount of random noise, a predictor dimension in the thousand, and a sample noise in the dozens. One of the most exciting areas to which microarray technology has been applied is the challenge of deciphering complex disease such as cancer. In these studies, samples are taken from two or more groups of individuals with heterogeneous phenotypes, pathologies, or clinical outcomes. these samples are hybridized to microarrays in an effort to find a small number of genes which are strongly correlated with the group of individuals. Eventhough today methods to analyse the data are welle developed and close to reach a standard organization (through the effort of preposed International project like Microarray Gene Expression Data -MGED- Society [1]) it is not unfrequant to stumble in a clinician's question that do not have a compelling statistical method that could permit to answer it.The contribution of this dissertation in deciphering disease regards the development of new approaches aiming at handle open problems posed by clinicians in handle specific experimental designs. In Chapter 1 starting from a biological necessary introduction, we revise the microarray tecnologies and all the important steps that involve an experiment from the production of the array, to the quality controls ending with preprocessing steps that will be used into the data analysis in the rest of the dissertation. While in Chapter 2 a critical review of standard analysis methods are provided stressing most of problems that In Chapter 3 is introduced a method to adress the issue of unbalanced design of miacroarray experiments. In microarray experiments, experimental design is a crucial starting-point for obtaining reasonable results. In a two-class problem, an equal or similar number of samples it should be collected between the two classes. However in some cases, e.g. rare pathologies, the approach to be taken is less evident. We propose to address this issue by applying a modified version of SAM [2]. MultiSAM consists in a reiterated application of a SAM analysis, comparing the less populated class (LPC) with 1,000 random samplings of the same size from the more populated class (MPC) A list of the differentially expressed genes is generated for each SAM application. After 1,000 reiterations, each single probe given a "score" ranging from 0 to 1,000 based on its recurrence in the 1,000 lists as differentially expressed. The performance of MultiSAM was compared to the performance of SAM and LIMMA [3] over two simulated data sets via beta and exponential distribution. The results of all three algorithms over low- noise data sets seems acceptable However, on a real unbalanced two-channel data set reagardin Chronic Lymphocitic Leukemia, LIMMA finds no significant probe, SAM finds 23 significantly changed probes but cannot separate the two classes, while MultiSAM finds 122 probes with score >300 and separates the data into two clusters by hierarchical clustering. We also report extra-assay validation in terms of differentially expressed genes Although standard algorithms perform well over low-noise simulated data sets, multi-SAM seems to be the only one able to reveal subtle differences in gene expression profiles on real unbalanced data. In Chapter 4 a method to adress similarities evaluation in a three-class prblem by means of Relevance Vector Machine [4] is described. In fact, looking at microarray data in a prognostic and diagnostic clinical framework, not only differences could have a crucial role. In some cases similarities can give useful and, sometimes even more, important information. The goal, given three classes, could be to establish, with a certain level of confidence, if the third one is similar to the first or the second one. In this work we show that Relevance Vector Machine (RVM) [2] could be a possible solutions to the limitation of standard supervised classification. In fact, RVM offers many advantages compared, for example, with his well-known precursor (Support Vector Machine - SVM [3]). Among these advantages, the estimate of posterior probability of class membership represents a key feature to address the similarity issue. This is a highly important, but often overlooked, option of any practical pattern recognition system. We focused on Tumor-Grade-three-class problem, so we have 67 samples of grade I (G1), 54 samples of grade 3 (G3) and 100 samples of grade 2 (G2). The goal is to find a model able to separate G1 from G3, then evaluate the third class G2 as test-set to obtain the probability for samples of G2 to be member of class G1 or class G3. The analysis showed that breast cancer samples of grade II have a molecular profile more similar to breast cancer samples of grade I. Looking at the literature this result have been guessed, but no measure of significance was gived before.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

[EN]The use of IT for teaching and learning is widely accepted as a means to enhance the learning experience. Hence, education professionals at all levels experience the impulse to introduce some kind of IT design in classrooms of every kind, where the use of IT has, at points, become mandatory. Nevertheless, there are little conclusive data that pinpoints what are the exact benefits that a given IT design, per se, brings to teaching or learning [1,2,3,4]. As any other technology, we contend, IT should be closely associated to the teaching methodology to be implemented, having into account all the factors that are going to influence all the process. In this article, we will analyse parameters that are considered to be critical if we are to predict the posible success of an IT design.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Nowadays, computing is migrating from traditional high performance and distributed computing to pervasive and utility computing based on heterogeneous networks and clients. The current trend suggests that future IT services will rely on distributed resources and on fast communication of heterogeneous contents. The success of this new range of services is directly linked to the effectiveness of the infrastructure in delivering them. The communication infrastructure will be the aggregation of different technologies even though the current trend suggests the emergence of single IP based transport service. Optical networking is a key technology to answer the increasing requests for dynamic bandwidth allocation and configure multiple topologies over the same physical layer infrastructure, optical networks today are still “far” from accessible from directly configure and offer network services and need to be enriched with more “user oriented” functionalities. However, current Control Plane architectures only facilitate efficient end-to-end connectivity provisioning and certainly cannot meet future network service requirements, e.g. the coordinated control of resources. The overall objective of this work is to provide the network with the improved usability and accessibility of the services provided by the Optical Network. More precisely, the definition of a service-oriented architecture is the enable technology to allow user applications to gain benefit of advanced services over an underlying dynamic optical layer. The definition of a service oriented networking architecture based on advanced optical network technologies facilitates users and applications access to abstracted levels of information regarding offered advanced network services. This thesis faces the problem to define a Service Oriented Architecture and its relevant building blocks, protocols and languages. In particular, this work has been focused on the use of the SIP protocol as a inter-layers signalling protocol which defines the Session Plane in conjunction with the Network Resource Description language. On the other hand, an advantage optical network must accommodate high data bandwidth with different granularities. Currently, two main technologies are emerging promoting the development of the future optical transport network, Optical Burst and Packet Switching. Both technologies respectively promise to provide all optical burst or packet switching instead of the current circuit switching. However, the electronic domain is still present in the scheduler forwarding and routing decision. Because of the high optics transmission frequency the burst or packet scheduler faces a difficult challenge, consequentially, high performance and time focused design of both memory and forwarding logic is need. This open issue has been faced in this thesis proposing an high efficiently implementation of burst and packet scheduler. The main novelty of the proposed implementation is that the scheduling problem has turned into simple calculation of a min/max function and the function complexity is almost independent of on the traffic conditions.