703 research outputs found

    Efficient Approaches for Voice Change and Voice Conversion Systems

    Get PDF
    In this thesis, the study and design of Voice Change and Voice Conversion systems are presented. Particularly, a voice change system manipulates a speaker’s voice to be perceived as it is not spoken by this speaker; and voice conversion system modifies a speaker’s voice, such that it is perceived as being spoken by a target speaker. This thesis mainly includes two sub-parts. The first part is to develop a low latency and low complexity voice change system (i.e. includes frequency/pitch scale modification and formant scale modification algorithms), which can be executed on the smartphones in 2012 with very limited computational capability. Although some low-complexity voice change algorithms have been proposed and studied, the real-time implementations are very rare. According to the experimental results, the proposed voice change system achieves the same quality as the baseline approach but requires much less computational complexity and satisfies the requirement of real-time. Moreover, the proposed system has been implemented in C language and was released as a commercial software application. The second part of this thesis is to investigate a novel low-complexity voice conversion system (i.e. from a source speaker A to a target speaker B) that improves the perceptual quality and identity without introducing large processing latencies. The proposed scheme directly manipulates the spectrum using an effective and physically motivated method – Continuous Frequency Warping and Magnitude Scaling (CFWMS) to guarantee high perceptual naturalness and quality. In addition, a trajectory limitation strategy is proposed to prevent the frame-by-frame discontinuity to further enhance the speech quality. The experimental results show that the proposed method outperforms the conventional baseline solutions in terms of either objective tests or subjective tests

    Modeling network traffic on a global network-centric system with artificial neural networks

    Get PDF
    This dissertation proposes a new methodology for modeling and predicting network traffic. It features an adaptive architecture based on artificial neural networks and is especially suited for large-scale, global, network-centric systems. Accurate characterization and prediction of network traffic is essential for network resource sizing and real-time network traffic management. As networks continue to increase in size and complexity, the task has become increasingly difficult and current methodology is not sufficiently adaptable or scaleable. Current methods model network traffic with express mathematical equations which are not easily maintained or adjusted. The accuracy of these models is based on detailed characterization of the traffic stream which is measured at points along the network where the data is often subject to constant variation and rapid evolution. The main contribution of this dissertation is development of a methodology that allows utilization of artificial neural networks with increased capability for adaptation and scalability. Application on an operating global, broadband network, the Connexion by Boeingʼ network, was evaluated to establish feasibility. A simulation model was constructed and testing was conducted with operational scenarios to demonstrate applicability on the case study network and to evaluate improvements in accuracy over existing methods --Abstract, page iii

    An Optimized Recursive General Regression Neural Network Oracle for the Prediction and Diagnosis of Diabetes

    Get PDF
    Diabetes is a serious, chronic disease that has been seeing a rise in the number of cases and prevalence over the past few decades. It can lead to serious complications and can increase the overall risk of dying prematurely. Data-oriented prediction models have become effective tools that help medical decision-making and diagnoses in which the use of machine learning in medicine has increased substantially. This research introduces the Recursive General Regression Neural Network Oracle (R-GRNN Oracle) and is applied on the Pima Indians Diabetes dataset for the prediction and diagnosis of diabetes. The R-GRNN Oracle (Bani-Hani, 2017) is an enhancement to the GRNN Oracle developed by Masters et al. in 1998, in which the recursive model is created of two oracles: one within the other. Several classifiers, along with the R-GRNN Oracle and the GRNN Oracle, are applied to the dataset, they are: Support Vector Machine (SVM), Multilayer Perceptron (MLP), Probabilistic Neural Network (PNN), Gaussian NaEF;ve Bayes (GNB), K-Nearest Neighbor (KNN), and Random Forest (RF). Genetic Algorithm (GA) was used for feature selection as well as the hyperparameter optimization of SVM and MLP, and Grid Search (GS) was used to optimize the hyperparameters of KNN and RF. The performance metrics accuracy, AUC, sensitivity, and specificity were recorded for each classifier. The R-GRNN Oracle was able to achieve the highest accuracy, AUC, and sensitivity (81.14%, 86.03%, and 63.80%, respectively), while the optimized MLP had the highest specificity (89.71%)

    Analyzing Cough Sounds for the Evidence of Covid-19 using Deep Learning Models

    Get PDF
    Early detection of infectious disease is the must to prevent/avoid multiple infections, and Covid-19 is an example. When dealing with Covid-19 pandemic, Cough is still ubiquitously presented as one of the key symptoms in both severe and non-severe Covid-19 infections, even though symptoms appear differently in different sociodemographic categories. By realizing the importance of clinical studies, analyzing cough sounds using AI-driven tools could help add more values when it comes to decision-making. Moreover, for mass screening and to serve resource constrained regions, AI-driven tools are the must. In this thesis, Convolutional Neural Network (CNN) tailored deep learning models are studied to analyze cough sounds to detect the possible evidence of Covid-19. In addition to custom CNN, pre-trained deep learning models (e.g., Vgg-16, Resnet-50, MobileNetV1, and DenseNet121) are employed on a publicly available dataset. In our findings, custom CNN performed comparatively better than pre-trained deep learning models

    Non-Gaussian Hybrid Transfer Functions: Memorizing Mine Survivability Calculations

    Get PDF
    Hybrid algorithms and models have received significant interest in recent years and are increasingly used to solve real-world problems. Different from existing methods in radial basis transfer function construction, this study proposes a novel nonlinear-weight hybrid algorithm involving the non-Gaussian type radial basis transfer functions. The speed and simplicity of the non-Gaussian type with the accuracy and simplicity of radial basis function are used to produce fast and accurate on-the-fly model for survivability of emergency mine rescue operations, that is, the survivability under all conditions is precalculated and used to train the neural network. The proposed hybrid uses genetic algorithm as a learning method which performs parameter optimization within an integrated analytic framework, to improve network efficiency. Finally, the network parameters including mean iteration, standard variation, standard deviation, convergent time, and optimized error are evaluated using the mean squared error. The results demonstrate that the hybrid model is able to reduce the computation complexity, increase the robustness and optimize its parameters. This novel hybrid model shows outstanding performance and is competitive over other existing models

    Evaluation of preprocessors for neural network speaker verification

    Get PDF

    Efficient audio signal processing for embedded systems

    Get PDF
    We investigated two design strategies that would allow us to efficiently process audio signals on embedded systems such as mobile phones and portable electronics. In the first strategy, we exploit properties of the human auditory system to process audio signals. We designed a sound enhancement algorithm to make piezoelectric loudspeakers sound "richer" and "fuller," using a combination of bass extension and dynamic range compression. We also developed an audio energy reduction algorithm for loudspeaker power management by suppressing signal energy below the masking threshold. In the second strategy, we use low-power analog circuits to process the signal before digitizing it. We designed an analog front-end for sound detection and implemented it on a field programmable analog array (FPAA). The sound classifier front-end can be used in a wide range of applications because programmable floating-gate transistors are employed to store classifier weights. Moreover, we incorporated a feature selection algorithm to simplify the analog front-end. A machine learning algorithm AdaBoost is used to select the most relevant features for a particular sound detection application. We also designed the circuits to implement the AdaBoost-based analog classifier.PhDCommittee Chair: Anderson, David; Committee Member: Hasler, Jennifer; Committee Member: Hunt, William; Committee Member: Lanterman, Aaron; Committee Member: Minch, Bradle

    CES-513 Stages for Developing Control Systems using EMG and EEG Signals: A survey

    Get PDF
    Bio-signals such as EMG (Electromyography), EEG (Electroencephalography), EOG (Electrooculogram), ECG (Electrocardiogram) have been deployed recently to develop control systems for improving the quality of life of disabled and elderly people. This technical report aims to review the current deployment of these state of the art control systems and explain some challenge issues. In particular, the stages for developing EMG and EEG based control systems are categorized, namely data acquisition, data segmentation, feature extraction, classification, and controller. Some related Bio-control applications are outlined. Finally a brief conclusion is summarized.

    Load forecast on a Micro Grid level through Machine Learning algorithms

    Get PDF
    As Micro Redes constituem um sector em crescimento da indústria energética, representando uma mudança de paradigma, desde as remotas centrais de geração até à produção mais localizada e distribuída. A capacidade de isolamento das principais redes elétricas e atuar de forma independente tornam as Micro Redes em sistemas resilientes, capazes de conduzir operações flexíveis em paralelo com a prestação de serviços que tornam a rede mais competitiva. Como tal, as Micro Redes fornecem energia limpa eficiente de baixo custo, aprimoram a coordenação dos ativos e melhoram a operação e estabilidade da rede regional de eletricidade, através da capacidade de resposta dinâmica aos recursos energéticos. Para isso, necessitam de uma coordenação de gestão inteligente que equilibre todas as tecnologias ao seu dispor. Daqui surge a necessidade de recorrer a modelos de previsão de carga e de produção robustos e de confiança, que interligam a alocação dos recursos da rede perante as necessidades emergentes. Sendo assim, foi desenvolvida a metodologia HALOFMI, que tem como principal objetivo a criação de um modelo de previsão de carga para 24 horas. A metodologia desenvolvida é constituída, numa primeira fase, por uma abordagem híbrida de multinível para a criação e escolha de atributos, que alimenta uma rede neuronal (Multi-Layer Perceptron) sujeita a um ajuste de híper-parâmetros. Posto isto, numa segunda fase são testados dois modos de aplicação e gestão de dados para a Micro Rede. A metodologia desenvolvida é aplicada em dois casos de estudo: o primeiro é composto por perfis de carga agregados correspondentes a dados de clientes em Baixa Tensão Normal e de Unidades de Produção e Autoconsumo (UPAC). Este caso de estudo apresenta-se como um perfil de carga elétrica regular e com contornos muito suaves. O segundo caso de estudo diz respeito a uma ilha turística e representa um perfil irregular de carga, com variações bruscas e difíceis de prever e apresenta um desafio maior em termos de previsão a 24-horas A partir dos resultados obtidos, é avaliado o impacto da integração de uma seleção recursiva inteligente de atributos, seguido por uma viabilização do processo de redução da dimensão de dados para o operador da Micro Rede, e por fim uma comparação de estimadores usados no modelo de previsão, através de medidores de erros na performance do algoritmo.Micro Grids constitute a growing sector of the energetic industry, representing a paradigm shift from the central power generation plans to a more distributed generation. The capacity to work isolated from the main electric grid make the MG resilient system, capable of conducting flexible operations while providing services that make the network more competitive. Additionally, Micro Grids supply clean and efficient low-cost energy, enhance the flexible assets coordination and improve the operation and stability of the of the local electric grid, through the capability of providing a dynamic response to the energetic resources. For that, it is required an intelligent coordination which balances all the available technologies. With this, rises the need to integrate accurate and robust load and production forecasting models into the MG management platform, thus allowing a more precise coordination of the flexible resource according to the emerging demand needs. For these reasons, the HALOFMI methodology was developed, which focus on the creation of a precise 24-hour load forecast model. This methodology includes firstly, a hybrid multi-level approach for the creation and selection of features. Then, these inputs are fed to a Neural Network (Multi-Layer Perceptron) with hyper-parameters tuning. In a second phase, two ways of data operation are compared and assessed, which results in the viability of the network operating with a reduced number of training days without compromising the model's performance. Such process is attained through a sliding window application. Furthermore, the developed methodology is applied in two case studies, both with 15-minute timesteps: the first one is composed by aggregated load profiles of Standard Low Voltage clients, including production and self-consumption units. This case study presents regular and very smooth load profile curves. The second case study concerns a touristic island and represents an irregular load curve with high granularity with abrupt variations. From the attained results, it is evaluated the impact of integrating a recursive intelligent feature selection routine, followed by an assessment on the sliding window application and at last, a comparison on the errors coming from different estimators for the model, through several well-defined performance metrics
    corecore