86 research outputs found
Seperation & Recovery of Cu and Zn by Solvent Extraction - Electrowinning from the Sulphate Leach Liquor of Complex Sulphide Ore
Complex sulphide ores containing Cu, Zn, and Pb form an important source of base metals in India. The sulphides are first floated out and the buck sulphide concentrates are subjected to roasting in fluidised bed reactor, where the metal sulphides are converted to setluble sulphates. The roasted mass is leached with water when copper and zinc get solubilised leaving lead in the insoluble residue. The solution is processed further by solvent extraction technique to extract copper quantitatively using the commercial extractant, LIX 64N, The zinc remains unextracted and can be recovered by electro-winning. This paper describes the process developed and the results obtained on recovery of electrolytic grade- copper and zinc
A model problem for conformal parameterizations of the Einstein constraint equations
We investigate the possibility that the conformal and conformal thin sandwich
(CTS) methods can be used to parameterize the set of solutions of the vacuum
Einstein constraint equations. To this end we develop a model problem obtained
by taking the quotient of certain symmetric data on conformally flat tori.
Specializing the model problem to a three-parameter family of conformal data we
observe a number of new phenomena for the conformal and CTS methods. Within
this family, we obtain a general existence theorem so long as the mean
curvature does not change sign. When the mean curvature changes sign, we find
that for certain data solutions exist if and only if the transverse-traceless
tensor is sufficiently small. When such solutions exist, there are generically
more than one. Moreover, the theory for mean curvatures changing sign is shown
to be extremely sensitive with respect to the value of a coupling constant in
the Einstein constraint equations.Comment: 40 pages, 4 figure
A process pattern model for tackling and improving big data quality
Data seldom create value by themselves. They need to be linked and combined from multiple sources, which can often come with variable data quality. The task of improving data quality is a recurring challenge. In this paper, we use a case study of a large telecom company to develop a generic process pattern model for improving data quality. The process pattern model is defined as a proven series of activities, aimed at improving the data quality given a certain context, a particular objective, and a specific set of initial conditions. Four different patterns are derived to deal with the variations in data quality of datasets. Instead of having to find the way to improve the quality of big data for each situation, the process model provides data users with generic patterns, which can be used as a reference model to improve big data quality
Effects of chronic inflammatory bowel diseases on left ventricular structure and function: a study protocol
BACKGROUND: Experimental evidences suggest an increased collagen deposition in inflammatory bowel diseases (IBD). In particular, large amounts of collagen type I, III and V have been described and correlated to the development of intestinal fibrotic lesions. No information has been available until now about the possible increased collagen deposition far from the main target organ. In the hypothesis that chronic inflammation and increased collagen metabolism are reflected also in the systemic circulation, we aimed this study to evaluate the effects on left ventricular wall structure by assessing splancnic and systemic collagen metabolism (procollagen III assay), deposition (ultrasonic tissue characterization), and cardiac function (echocardiography) in patients with different long standing history of IBD, before and after surgery. METHODS: Thirty patients affected by active IBD, 15 with Crohn and 15 with Ulcerative Colitis, submitted to surgery will be enrolled in the study in a double blind fashion. They will be studied before the surgical operation and 6, 12 months after surgery. A control group of 15 healthy age and gender-matched subjects will also be studied. At each interval blood samples will be collected in order to assess the collagen metabolism; a transthoracic echocardiogram will be recorded for the subsequent determination of cardiac function and collagen deposition. DISCUSSION: From this study protocol we expect additional information about the association between IBD and cardiovascular disorders; in particular to address the question if chronic inflammation, through the altered collagen metabolism, could affect left ventricular structure and function in a manner directly related to the estimated duration of the disease
Feasibility study of computed tomography colonography using limited bowel preparation at normal and low-dose levels study
The purpose was to evaluate low-dose CT colonography without cathartic cleansing in terms of image quality, polyp visualization and patient acceptance. Sixty-one patients scheduled for colonoscopy started a low-fiber diet, lactulose and amidotrizoic-acid for fecal tagging 2 days prior to the CT scan (standard dose, 5.8–8.2 mSv). The original raw data of 51 patients were modified and reconstructed at simulated 2.3 and 0.7 mSv levels. Two observers evaluated the standard dose scan regarding image quality and polyps. A third evaluated the presence of polyps at all three mSv levels in a blinded prospective way. All observers were blinded to the reference standard: colonoscopy. At three times patients were given questionnaires relating to their experiences and preference. Image quality was sufficient in all patients, but significantly lower in the cecum, sigmoid and rectum. The two observers correctly identified respectively 10/15 (67%) and 9/15 (60%) polyps ≥10 mm, with 5 and 8 false-positive lesions (standard dose scan). Dose reduction down to 0.7 mSv was not associated with significant changes in diagnostic value (polyps ≥10 mm). Eighty percent of patients preferred CT colonography and 13% preferred colonoscopy (P<0.001). CT colonography without cleansing is preferred to colonoscopy and shows sufficient image quality and moderate sensitivity, without impaired diagnostic value at dose-levels as low as 0.7 mSv
Visualizing Big Data with augmented and virtual reality: challenges and research agenda
This paper provides a multi-disciplinary overview of the research issues and achievements in the field of Big Data and its visualization techniques and tools. The main aim is to summarize challenges in visualization methods for existing Big Data, as well as to offer novel solutions for issues related to the current state of Big Data Visualization. This paper provides a classification of existing data types, analytical methods, visualization techniques and tools, with a particular emphasis placed on surveying the evolution of visualization methodology over the past years. Based on the results, we reveal disadvantages of existing visualization methods. Despite the technological development of the modern world, human involvement (interaction), judgment and logical thinking are necessary while working with Big Data. Therefore, the role of human perceptional limitations involving large amounts of information is evaluated. Based on the results, a non-traditional approach is proposed: we discuss how the capabilities of Augmented Reality and Virtual Reality could be applied to the field of Big Data Visualization. We discuss the promising utility of Mixed Reality technology integration with applications in Big Data Visualization. Placing the most essential data in the central area of the human visual field in Mixed Reality would allow one to obtain the presented information in a short period of time without significant data losses due to human perceptual issues. Furthermore, we discuss the impacts of new technologies, such as Virtual Reality displays and Augmented Reality helmets on the Big Data visualization as well as to the classification of the main challenges of integrating the technology.publishedVersionPeer reviewe
Recommended from our members
Using Robson's Ten‐Group Classification System for comparing caesarean section rates in Europe: an analysis of routine data from the Euro‐Peristat study
Objective
Robson's Ten Group Classification System (TGCS) creates clinically relevant sub‐groups for monitoring caesarean birth rates. This study assesses whether this classification can be derived from routine data in Europe and uses it to analyse national caesarean rates.
Design
Observational study using routine data.
Setting
Twenty‐seven EU member states plus Iceland, Norway, Switzerland and the UK.
Population
All births at ≥22 weeks of gestational age in 2015.
Methods
National statistical offices and medical birth registers derived numbers of caesarean births in TGCS groups.
Main outcome measures
Overall caesarean rate, prevalence and caesarean rates in each of the TGCS groups.
Results
Of 31 countries, 18 were able to provide data on the TGCS groups, with UK data available only from Northern Ireland. Caesarean birth rates ranged from 16.1 to 56.9%. Countries providing TGCS data had lower caesarean rates than countries without data (25.8% versus 32.9%, P = 0.04). Countries with higher caesarean rates tended to have higher rates in all TGCS groups. Substantial heterogeneity was observed, however, especially for groups 5 (previous caesarean section), 6, 7 (nulliparous/multiparous breech) and 10 (singleton cephalic preterm). The differences in percentages of abnormal lies, group 9, illustrate potential misclassification arising from unstandardised definitions.
Conclusions
Although further validation of data quality is needed, using TGCS in Europe provides valuable comparator and baseline data for benchmarking and surveillance. Higher caesarean rates in countries unable to construct the TGCS suggest that effective routine information systems may be an indicator of a country's investment in implementing evidence‐based caesarean policies.
Tweetable abstract
Many European countries can provide Robson's Ten‐Group Classification to improve caesarean rate comparisons
Normal parameter reduction algorithm in soft set based on hybrid binary particle swarm and biogeography optimizer
© 2019, Springer-Verlag London Ltd., part of Springer Nature. Existing classification techniques that are proposed previously for eliminating data inconsistency could not achieve an efficient parameter reduction in soft set theory, which effects on the obtained decisions. Meanwhile, the computational cost made during combination generation process of soft sets could cause machine infinite state, which is known as nondeterministic polynomial time. The contributions of this study are mainly focused on minimizing choices costs through adjusting the original classifications by decision partition order and enhancing the probability of searching domain space using a developed Markov chain model. Furthermore, this study introduces an efficient soft set reduction-based binary particle swarm optimized by biogeography-based optimizer (SSR-BPSO-BBO) algorithm that generates an accurate decision for optimal and sub-optimal choices. The results show that the decision partition order technique is performing better in parameter reduction up to 50%, while other algorithms could not obtain high reduction rates in some scenarios. In terms of accuracy, the proposed SSR-BPSO-BBO algorithm outperforms the other optimization algorithms in achieving high accuracy percentage of a given soft dataset. On the other hand, the proposed Markov chain model could significantly represent the robustness of our parameter reduction technique in obtaining the optimal decision and minimizing the search domain.Published versio
- …