609 research outputs found
تمثيل الإطار الخارجي للكلمات العربية بكفاءة من خلال الدمج بين نموذج الكنتور النشط وتحديد ونقاط الزوايا
Graphical curves and surfaces fitting are hot areas of research studies and application, such as artistic applications, analysis applications and encoding purposes. Outline capture of digital word images is important in most of the desktop publishing systems. The shapes of the characters are stored in the computer memory in terms of their outlines, and the outlines are expressed as Bezier curves. Existing methods for Arabic font outline description suffer from low fitting accuracy and efficiency. In our research, we developed a new method for outlining shapes using Bezier curves with minimal set of curve points. A distinguishing characteristic of our method is that it combines the active contour method (snake) with corner detection to achieve an initial set of points that is as close to the shape's boundaries as possible. The method links these points (snake + corner) into a compound Bezier curve, and iteratively improves the fitting of the curve over the actual boundaries of the shape. We implemented and tested our method using MATLAB. Test cases included various levels of shape complexity varying from simple, moderate, and high complexity depending on factors, such as: boundary concavities, number of corners. Results show that our method achieved average 86% of accuracy when measured relative to true shape boundary. When compared to other similar methods (Masood & Sarfraz, 2009; Sarfraz & Khan, 2002; Ferdous A Sohel, Karmakar, Dooley, & Bennamoun, 2010), our method performed comparatively well. Keywords: Bezier curves, shape descriptor, curvature, corner points, control points, Active Contour Model.تعتبر المنحنيات والأسطح الرسومية موضوعاً هاماً في الدراسات البحثية وفي التطبيقات البرمجية مثل التطبيقات الفنية، وتطبيقات تحليل وترميز البيانات. ويعتبر تخطيط الحدود الخارجية للكلمات عملية أساسية في غالبية تطبيقات النشر المكتبي. في هذه التطبيقات تخزن أشكال الأحرف في الذاكرة من حيث خطوطها الخارجية، وتمثل الخطوط الخارجية على هيئة منحنيات Bezier. الطرق المستخدمة حالياً لتحديد الخطوط الخارجية للكلمات العربية تنقصها دقة وكفاءة الملاءمة ما بين الحدود الحقيقية والمنحنى الرسومي الذي تقوم بتشكيله. في هذا البحث قمنا بتطوير طريقة جديدة لتخطيط الحدود الخارجية للكلمات تعتمد على منحنيات Bezier بمجموعة أقل من المنحنيات الجزئية. تتميز طريقتنا بخاصية مميزة وهي الدمج بين آلية لاستشعار الزوايا مع آلية نموذج الكنتور النشط (الأفعى). يتم الدمج بين نقاط الزوايا ونقاط الأفعى لتشكيل مجموعة موحدة من النقاط المبدئية قريبة قدر الإمكان من الحدود الحقيقية للشكل المراد تحديده. يتشكل منحنى Bezier من هذه المجموعة المدمجة، وتتم عملية تدريجية على دورات لملاءمة المنحنى على الحدود الحقيقية للشكل. قام الباحث بتنفيذ وتجربة الطريقة الجديدة باستخدام برنامج MATLAB. وتم اختيار أشكال رسومية كعينات اختبار تتصف بمستويات متباينة من التعقيد تتراوح ما بين بسيط إلى متوسط إلى عالي التعقيد على أساس عوامل مثل تقعرات الحدود، عدد نقاط الزوايا، الفتحات الداخلية، إلخ. وقد أظهرت نتائج الاختبار أن طريقتنا الجديدة حققت دقة في الملائمة تصل نسبتها إلى 86% مقارنة بالحدود الحقيقية للشكل المستهدف. وكذلك فقد كان أداء طريقتنا جيداً بالمقارنة مع طرق أخرى مماثلة
Recommended from our members
On the capture and representation of fonts
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The commercial need to capture, process and represent the shape and form of an outline has lead to the development of a number of spline routines. These use a mathematical curve format that approximates the contours of a given shape. The modelled outline lends itself to be used on, and for, a variety of purposes. These include graphic screens, laser printers and numerically controlled machines. The latter can be employed for cutting foil, metal. plastic and stone. One of the most widely used software design packages has been the lKARUS system. This, developed by URW of Hamburg (Gennany), employs a number of mathematical descriptions that facilitate the process of both modelling and representation of font characters. It uses a variety of curve formats, including Bezier cubics, general conics and parabolics. The work reported in this dissertation focuses on developing improved techniques, primarily. for the lKARUS system. This includes two algorithms
which allow a Bezier cubic description, two for a general conic representation and, yet another, two for the parabolic case. In addition, a number of algorithms are presented which promote conversions between these mathematical forms; for example, Bezier cubics to a general conic form. Furthennore, algorithms are developed to assist the process of rasterising both cubic and quadratic arcs.This study was partly funded by the Science and Education Research Council (SERC)
The development of a finite elements based springback compensation tool for sheet metal products
Springback is a major problem in the deep drawing process. When the tools are released after the forming stage, the product springs back due to the action of internal stresses. In many cases the shape deviation is too large and springback compensation is needed: the tools of the deep drawing process are changed so, that the product becomes geometrically accurate after springback. In this paper, two different ways of geometric optimization are presented, the smooth displacement adjustment (SDA) method and the surface controlled overbending (SCO) method. Both methods use results from a finite elements deep drawing simulation for the optimization of the tool shape. The methods are demonstrated on an industrial product. The results are satisfactory, but it is shown that both methods still need to be improved and that the FE simulation needs to become more reliable to allow industrial application
08221 Abstracts Collection -- Geometric Modeling
From May 26 to May 30 2008 the Dagstuhl Seminar 08221 ``Geometric Modeling\u27\u27 was held in the International Conference and Research Center (IBFI),
Schloss Dagstuhl.
During the seminar, several participants presented their current
research, and ongoing work and open problems were discussed. Abstracts of
the presentations given during the seminar as well as abstracts of
seminar results and ideas are put together in this paper. The first section
describes the seminar topics and goals in general.
Links to extended abstracts or full papers are provided, if available
Shape analysis of the corpus callosum of autistic and normal subjects in neuroimaging.
Early detection of human disease in today’s society can have an enormous impact on the severity of the disease that is manifested. Disease such as Autism and Dyslexia, which have no current cure or proven mechanism as to how they develop, can often have an adverse physical and physiological impact on the lifestyle of a human being. Although these disease are not fully curable, the severity handicaps that accompany them can be significantly reduced with the proper therapy, and thus the earlier that the disease is detected the faster therapy can be administered. The research in this thesis is an attempt at studying discriminatory shape measures of some brain structures that are known to carry changes from autistics to normal individuals. The focus will be on the corpus callosum. There has been considerable research done on the brain scans (MRI, CT) of autistic individuals vs. control (normal) individuals to observe any noticeable discrepancies through statistical analysis. The most common and powerful tool to analyze structures of the brain, once a specific region has been segmented, is using Registration to match like structures and record their error. The ICP algorithm (Iterative Closest Point) is commonly used to accomplish this task. Many techniques such as level sets and statistical methods can be used for segmentation. The Corpus Callosum (CC) and the cortical surface of the brain are currently where most Autism analysis is performed. It has been observed that the gyrification of the cortical surface is different in the two groups, and size as well as shape of the CC. An analysis approach for autism MRI is quite extensive and involves many steps. This thesis is limited to examination of shape measures of the CC that lend discrimination ability to distinguish between normal and autistic individuals from T1-weigheted MRI scans. We will examine two approaches for shape analysis, based on the traditional Fourier Descriptors (FD) method and shape registration (SR) using the procrustes technique. MRI scans of 22 autistic and 16 normal individuals are used to test the approaches developed in this thesis. We show that both FD and SR may be used to extract features to discriminate between the two populations with accuracy levels over 80% up to 100% depending on the technique
Recommended from our members
The statistical properties and coding of handwriting and hand-drawn graphics in tutorial classes
This research investigates compression techniques as applied to input from a digitizing tablet. The representations we consider are either drawn from the literature or developed within this thesis; they involve selecting relevant points along the trajectory of the writing pen, and using an interpolating function for reconstruction. The two dimensional parametric polynomial interpolator (e.g linear, quadratic, cubic ) is used for regenerating the trajectory of the pen. Several techniques for selecting relevant points , i.e control points are discussed. We determine our best representation by weighting the criteria, analyzing the experimental results, examining the theoretical considerations and, making where necessary, justifiable tradeoffs. The techniques tested are objectively evaluated in terms of accuracy, efficiency and compactness. The information characteristics (i.e signal entropy) of the handwriting and drawing signals, are estimated from the original and approximated data. Results are discussed and conclusions drawn. The research has been carried out on static data
A generic shape descriptor using Bezier curves
Bezier curves are robust tool for a wide array of applications ranging from computer-aided design to calligraphic character, outlining and object shape description. In terms of the control point generation process, existing shape descriptor techniques that employ Bezier curves do not distinguish between regions where an object's shape changes rapidly and those where the change is more gradual or flat. This can lead to an erroneous shape description, particularly where there are significantly sharp changes in shape, such as at sharp corners. This paper presents a novel shape description algorithm called a generic shape descriptor using Bezier curves (SDBC), which defines a new strategy for Bezier control point generation by integrating domain specific information about the shape of an object in a particular region. The strategy also includes an improved dynamic fixed length coding scheme for control points. The SDBC framework has been rigorously tested upon a number of arbitrary shapes, and both quantitative and qualitative analyses have confirmed its superior performance in comparison with existing algorithms
Component Segmentation of Engineering Drawings Using Graph Convolutional Networks
We present a data-driven framework to automate the vectorization and machine
interpretation of 2D engineering part drawings. In industrial settings, most
manufacturing engineers still rely on manual reads to identify the topological
and manufacturing requirements from drawings submitted by designers. The
interpretation process is laborious and time-consuming, which severely inhibits
the efficiency of part quotation and manufacturing tasks. While recent advances
in image-based computer vision methods have demonstrated great potential in
interpreting natural images through semantic segmentation approaches, the
application of such methods in parsing engineering technical drawings into
semantically accurate components remains a significant challenge. The severe
pixel sparsity in engineering drawings also restricts the effective
featurization of image-based data-driven methods. To overcome these challenges,
we propose a deep learning based framework that predicts the semantic type of
each vectorized component. Taking a raster image as input, we vectorize all
components through thinning, stroke tracing, and cubic bezier fitting. Then a
graph of such components is generated based on the connectivity between the
components. Finally, a graph convolutional neural network is trained on this
graph data to identify the semantic type of each component. We test our
framework in the context of semantic segmentation of text, dimension and,
contour components in engineering drawings. Results show that our method yields
the best performance compared to recent image, and graph-based segmentation
methods.Comment: Preprint accepted to Computers in Industr
- …