93 research outputs found
Fundamental activity constraints lead to specific interpretations of the connectome
The continuous integration of experimental data into coherent models of the
brain is an increasing challenge of modern neuroscience. Such models provide a
bridge between structure and activity, and identify the mechanisms giving rise
to experimental observations. Nevertheless, structurally realistic network
models of spiking neurons are necessarily underconstrained even if experimental
data on brain connectivity are incorporated to the best of our knowledge.
Guided by physiological observations, any model must therefore explore the
parameter ranges within the uncertainty of the data. Based on simulation
results alone, however, the mechanisms underlying stable and physiologically
realistic activity often remain obscure. We here employ a mean-field reduction
of the dynamics, which allows us to include activity constraints into the
process of model construction. We shape the phase space of a multi-scale
network model of the vision-related areas of macaque cortex by systematically
refining its connectivity. Fundamental constraints on the activity, i.e.,
prohibiting quiescence and requiring global stability, prove sufficient to
obtain realistic layer- and area-specific activity. Only small adaptations of
the structure are required, showing that the network operates close to an
instability. The procedure identifies components of the network critical to its
collective dynamics and creates hypotheses for structural data and future
experiments. The method can be applied to networks involving any neuron model
with a known gain function.Comment: J. Schuecker and M. Schmidt contributed equally to this wor
Application of Graph Neural Networks and graph descriptors for graph classification
Graph classification is an important area in both modern research and
industry. Multiple applications, especially in chemistry and novel drug
discovery, encourage rapid development of machine learning models in this area.
To keep up with the pace of new research, proper experimental design, fair
evaluation, and independent benchmarks are essential. Design of strong
baselines is an indispensable element of such works.
In this thesis, we explore multiple approaches to graph classification. We
focus on Graph Neural Networks (GNNs), which emerged as a de facto standard
deep learning technique for graph representation learning. Classical
approaches, such as graph descriptors and molecular fingerprints, are also
addressed. We design fair evaluation experimental protocol and choose proper
datasets collection. This allows us to perform numerous experiments and
rigorously analyze modern approaches. We arrive to many conclusions, which shed
new light on performance and quality of novel algorithms.
We investigate application of Jumping Knowledge GNN architecture to graph
classification, which proves to be an efficient tool for improving base graph
neural network architectures. Multiple improvements to baseline models are also
proposed and experimentally verified, which constitutes an important
contribution to the field of fair model comparison.Comment: Master's thesis submitted at AGH University of Science and Technolog
Data efficient deep learning models for text classification
Text classification is one of the most important techniques within natural language processing. Applications range from topic detection and intent identification to sentiment analysis. Usually text classification is formulated
as a supervised learning problem, where a labeled training set is fed into a machine learning algorithm. In practice, training supervised machine learning algorithms such as those present in deep learning, require
large training sets which involves a considerable amount of human labor to manually tag the data. This constitutes a bottleneck in applied supervised learning, and as a result, it is desired to have supervised learning models that require smaller amounts of tagged data.
In this work, we will research and compare supervised learning models for text classification that are data efficient, that is, require small amounts of tagged data to achieve state of the art performance levels. In particular,
we will study transfer learning techniques that reuse previous knowledge to train supervised learning models. For the purpose of comparison, we will focus on opinion polarity classification, a sub problem within sentiment
analysis that assigns polarity to an opinion (positive or negative) depending on the mood of the opinion holder.
Multiple deep learning models to learn representations of texts including BERT, InferSent, Universal Sentence Encoder and the Sentiment Neuron are compared in six datasets from different domains. Results show that transfer learning dramatically improves data efficiency, obtaining double digit improvements in F1 score just with under 100 supervised training examples
English/Russian lexical cognates detection using NLP Machine Learning with Python
ΠΠ·ΡΡΠ΅Π½ΠΈΠ΅ ΡΠ·ΡΠΊΠ° β ΡΡΠΎ Π·Π°ΠΌΠ΅ΡΠ°ΡΠ΅Π»ΡΠ½ΠΎΠ΅ Π·Π°Π½ΡΡΠΈΠ΅, ΠΊΠΎΡΠΎΡΠΎΠ΅ ΡΠ°ΡΡΠΈΡΡΠ΅Ρ Π½Π°Ρ ΠΊΡΡΠ³ΠΎΠ·ΠΎΡ ΠΈ ΠΏΠΎΠ·Π²ΠΎΠ»ΡΠ΅Ρ Π½Π°ΠΌ ΠΎΠ±ΡΠ°ΡΡΡΡ Ρ ΠΏΡΠ΅Π΄ΡΡΠ°Π²ΠΈΡΠ΅Π»ΡΠΌΠΈ ΡΠ°Π·Π»ΠΈΡΠ½ΡΡ
ΠΊΡΠ»ΡΡΡΡ ΠΈ Π»ΡΠ΄Π΅ΠΉ ΠΏΠΎ Π²ΡΠ΅ΠΌΡ ΠΌΠΈΡΡ. Π’ΡΠ°Π΄ΠΈΡΠΈΠΎΠ½Π½ΠΎ ΡΠ·ΡΠΊΠΎΠ²ΠΎΠ΅ ΠΎΠ±ΡΠ°Π·ΠΎΠ²Π°Π½ΠΈΠ΅ ΠΎΡΠ½ΠΎΠ²ΡΠ²Π°Π»ΠΎΡΡ Π½Π° ΡΡΠ°Π΄ΠΈΡΠΈΠΎΠ½Π½ΡΡ
ΠΌΠ΅ΡΠΎΠ΄Π°Ρ
, ΡΠ°ΠΊΠΈΡ
ΠΊΠ°ΠΊ ΡΡΠ΅Π±Π½ΠΈΠΊΠΈ, ΡΠ»ΠΎΠ²Π°ΡΠ½ΡΠΉ Π·Π°ΠΏΠ°Ρ ΠΈ ΡΠ·ΡΠΊΠΎΠ²ΠΎΠΉ ΠΎΠ±ΠΌΠ΅Π½. ΠΠ΄Π½Π°ΠΊΠΎ Ρ ΠΏΠΎΡΠ²Π»Π΅Π½ΠΈΠ΅ΠΌ ΠΌΠ°ΡΠΈΠ½Π½ΠΎΠ³ΠΎ ΠΎΠ±ΡΡΠ΅Π½ΠΈΡ Π½Π°ΡΡΡΠΏΠΈΠ»Π° Π½ΠΎΠ²Π°Ρ ΡΡΠ° Π² ΠΎΠ±ΡΡΠ΅Π½ΠΈΠΈ ΡΠ·ΡΠΊΡ, ΠΏΡΠ΅Π΄Π»Π°Π³Π°ΡΡΠ°Ρ ΠΈΠ½Π½ΠΎΠ²Π°ΡΠΈΠΎΠ½Π½ΡΠ΅ ΠΈ ΡΡΡΠ΅ΠΊΡΠΈΠ²Π½ΡΠ΅ ΡΠΏΠΎΡΠΎΠ±Ρ ΡΡΠΊΠΎΡΠ΅Π½ΠΈΡ ΠΎΠ²Π»Π°Π΄Π΅Π½ΠΈΡ ΡΠ·ΡΠΊΠΎΠΌ. ΠΠ΄Π½ΠΈΠΌ ΠΈΠ· ΠΈΠ½ΡΡΠΈΠ³ΡΡΡΠΈΡ
ΠΏΡΠΈΠΌΠ΅Π½Π΅Π½ΠΈΠΉ ΠΌΠ°ΡΠΈΠ½Π½ΠΎΠ³ΠΎ ΠΎΠ±ΡΡΠ΅Π½ΠΈΡ Π² ΠΈΠ·ΡΡΠ΅Π½ΠΈΠΈ ΡΠ·ΡΠΊΠΎΠ² ΡΠ²Π»ΡΠ΅ΡΡΡ ΠΈΡΠΏΠΎΠ»ΡΠ·ΠΎΠ²Π°Π½ΠΈΠ΅ ΡΠΎΠ΄ΡΡΠ²Π΅Π½Π½ΡΡ
ΡΠ»ΠΎΠ², ΡΠ»ΠΎΠ², ΠΊΠΎΡΠΎΡΡΠ΅ ΠΈΠΌΠ΅ΡΡ ΡΡ
ΠΎΠΆΠ΅Π΅ Π·Π½Π°ΡΠ΅Π½ΠΈΠ΅ ΠΈ Π½Π°ΠΏΠΈΡΠ°Π½ΠΈΠ΅ Π² ΡΠ°Π·Π½ΡΡ
ΡΠ·ΡΠΊΠ°Ρ
. ΠΠ»Ρ ΡΠ΅ΡΠ΅Π½ΠΈΡ ΡΡΠΎΠΉ ΡΠ΅ΠΌΡ Π² Π΄Π°Π½Π½ΠΎΠΉ ΠΈΡΡΠ»Π΅Π΄ΠΎΠ²Π°ΡΠ΅Π»ΡΡΠΊΠΎΠΉ ΡΠ°Π±ΠΎΡΠ΅ ΠΏΡΠ΅Π΄Π»Π°Π³Π°Π΅ΡΡΡ ΠΎΠ±Π»Π΅Π³ΡΠΈΡΡ ΠΏΡΠΎΡΠ΅ΡΡ ΠΈΠ·ΡΡΠ΅Π½ΠΈΡ Π²ΡΠΎΡΠΎΠ³ΠΎ ΡΠ·ΡΠΊΠ° Ρ ΠΏΠΎΠΌΠΎΡΡΡ ΠΈΡΠΊΡΡΡΡΠ²Π΅Π½Π½ΠΎΠ³ΠΎ ΠΈΠ½ΡΠ΅Π»Π»Π΅ΠΊΡΠ°, Π² ΡΠ°ΡΡΠ½ΠΎΡΡΠΈ Π½Π΅ΠΉΡΠΎΠ½Π½ΡΡ
ΡΠ΅ΡΠ΅ΠΉ, ΠΊΠΎΡΠΎΡΡΠ΅ ΠΌΠΎΠ³ΡΡ ΠΈΠ΄Π΅Π½ΡΠΈΡΠΈΡΠΈΡΠΎΠ²Π°ΡΡ ΠΈ ΠΈΡΠΏΠΎΠ»ΡΠ·ΠΎΠ²Π°ΡΡ ΡΠ»ΠΎΠ²Π°, ΠΏΠΎΡ
ΠΎΠΆΠΈΠ΅ ΠΈΠ»ΠΈ ΠΈΠ΄Π΅Π½ΡΠΈΡΠ½ΡΠ΅ ΠΊΠ°ΠΊ Π½Π° ΠΏΠ΅ΡΠ²ΠΎΠΌ ΡΠ·ΡΠΊΠ΅ ΡΡΠ°ΡΠ΅Π³ΠΎΡΡ, ΡΠ°ΠΊ ΠΈ Π½Π° ΡΠ΅Π»Π΅Π²ΠΎΠΌ ΡΠ·ΡΠΊΠ΅. ΠΡΠΈ ΡΠ»ΠΎΠ²Π°, ΠΈΠ·Π²Π΅ΡΡΠ½ΡΠ΅ ΠΊΠ°ΠΊ Π»Π΅ΠΊΡΠΈΡΠ΅ΡΠΊΠΈΠ΅ ΡΠΎΠ΄ΡΡΠ²Π΅Π½Π½ΡΠ΅ ΡΠ»ΠΎΠ²Π°, ΠΌΠΎΠ³ΡΡ ΠΎΠ±Π»Π΅Π³ΡΠΈΡΡ ΠΈΠ·ΡΡΠ΅Π½ΠΈΠ΅ ΡΠ·ΡΠΊΠ°, ΠΏΡΠ΅Π΄ΠΎΡΡΠ°Π²Π»ΡΡ ΡΡΠ°ΡΠΈΠΌΡΡ Π·Π½Π°ΠΊΠΎΠΌΡΠΉ ΠΎΡΠΈΠ΅Π½ΡΠΈΡ ΠΈ ΠΏΠΎΠ·Π²ΠΎΠ»ΡΡ ΠΈΠΌ ΡΠ²ΡΠ·ΡΠ²Π°ΡΡ Π½ΠΎΠ²ΡΠΉ ΡΠ»ΠΎΠ²Π°ΡΠ½ΡΠΉ Π·Π°ΠΏΠ°Ρ ΡΠΎ ΡΠ»ΠΎΠ²Π°ΠΌΠΈ, ΠΊΠΎΡΠΎΡΡΠ΅ ΠΎΠ½ΠΈ ΡΠΆΠ΅ Π·Π½Π°ΡΡ. ΠΡΠΏΠΎΠ»ΡΠ·ΡΡ Π²ΠΎΠ·ΠΌΠΎΠΆΠ½ΠΎΡΡΠΈ Π½Π΅ΠΉΡΠΎΠ½Π½ΡΡ
ΡΠ΅ΡΠ΅ΠΉ Π΄Π»Ρ ΠΎΠ±Π½Π°ΡΡΠΆΠ΅Π½ΠΈΡ ΠΈ ΠΈΡΠΏΠΎΠ»ΡΠ·ΠΎΠ²Π°Π½ΠΈΡ ΡΡΠΈΡ
ΡΠΎΠ΄ΡΡΠ²Π΅Π½Π½ΡΡ
ΡΠ»ΠΎΠ², ΡΡΠ°ΡΠΈΠ΅ΡΡ ΡΠΌΠΎΠ³ΡΡ ΡΡΠΊΠΎΡΠΈΡΡ ΡΠ²ΠΎΠΉ ΠΏΡΠΎΠ³ΡΠ΅ΡΡ Π² ΠΎΡΠ²ΠΎΠ΅Π½ΠΈΠΈ Π²ΡΠΎΡΠΎΠ³ΠΎ ΡΠ·ΡΠΊΠ°. Π₯ΠΎΡΡ ΠΈΡΡΠ»Π΅Π΄ΠΎΠ²Π°Π½ΠΈΠ΅ ΡΠ΅ΠΌΠ°Π½ΡΠΈΡΠ΅ΡΠΊΠΎΠ³ΠΎ ΡΡ
ΠΎΠ΄ΡΡΠ²Π° Π² ΡΠ°Π·Π½ΡΡ
ΡΠ·ΡΠΊΠ°Ρ
Π½Π΅ ΡΠ²Π»ΡΠ΅ΡΡΡ Π½ΠΎΠ²ΠΎΠΉ ΡΠ΅ΠΌΠΎΠΉ, Π½Π°ΡΠ° ΡΠ΅Π»Ρ ΡΠΎΡΡΠΎΠΈΡ Π² ΡΠΎΠΌ, ΡΡΠΎΠ±Ρ ΠΏΡΠΈΠΌΠ΅Π½ΠΈΡΡ Π΄ΡΡΠ³ΠΎΠΉ ΠΏΠΎΠ΄Ρ
ΠΎΠ΄ Π΄Π»Ρ Π²ΡΡΠ²Π»Π΅Π½ΠΈΡ ΡΡΡΡΠΊΠΎ-Π°Π½Π³Π»ΠΈΠΉΡΠΊΠΈΡ
Π»Π΅ΠΊΡΠΈΡΠ΅ΡΠΊΠΈΡ
ΡΠΎΠ΄ΡΡΠ²Π΅Π½Π½ΡΡ
ΡΠ»ΠΎΠ² ΠΈ ΠΏΡΠ΅Π΄ΡΡΠ°Π²ΠΈΡΡ ΠΏΠΎΠ»ΡΡΠ΅Π½Π½ΡΠ΅ ΡΠ΅Π·ΡΠ»ΡΡΠ°ΡΡ Π² ΠΊΠ°ΡΠ΅ΡΡΠ²Π΅ ΠΈΠ½ΡΡΡΡΠΌΠ΅Π½ΡΠ° ΠΈΠ·ΡΡΠ΅Π½ΠΈΡ ΡΠ·ΡΠΊΠ°, ΠΈΡΠΏΠΎΠ»ΡΠ·ΡΡ Π²ΡΠ±ΠΎΡΠΊΡ Π΄Π°Π½Π½ΡΡ
ΠΎ Π»Π΅ΠΊΡΠΈΡΠ΅ΡΠΊΠΎΠΌ ΠΈ ΡΠ΅ΠΌΠ°Π½ΡΠΈΡΠ΅ΡΠΊΠΎΠΌ ΡΡ
ΠΎΠ΄ΡΡΠ²Π΅. ΠΌΠ΅ΠΆΠ΄Ρ ΡΠ·ΡΠΊΠ°ΠΌΠΈ, ΡΡΠΎΠ±Ρ ΠΏΠΎΡΡΡΠΎΠΈΡΡ ΠΌΠΎΠ΄Π΅Π»Ρ ΠΎΠ±Π½Π°ΡΡΠΆΠ΅Π½ΠΈΡ Π»Π΅ΠΊΡΠΈΡΠ΅ΡΠΊΠΈΡ
ΡΠΎΠ΄ΡΡΠ²Π΅Π½Π½ΡΡ
ΡΠ»ΠΎΠ² ΠΈ Π°ΡΡΠΎΡΠΈΠ°ΡΠΈΠΉ ΡΠ»ΠΎΠ². ΠΠΏΠΎΡΠ»Π΅Π΄ΡΡΠ²ΠΈΠΈ, Π² Π·Π°Π²ΠΈΡΠΈΠΌΠΎΡΡΠΈ ΠΎΡ Π½Π°ΡΠ΅Π³ΠΎ Π°Π½Π°Π»ΠΈΠ·Π° ΠΈ ΡΠ΅Π·ΡΠ»ΡΡΠ°ΡΠΎΠ², ΠΌΡ ΠΏΡΠ΅Π΄ΡΡΠ°Π²ΠΈΠΌ ΠΏΡΠΈΠ»ΠΎΠΆΠ΅Π½ΠΈΠ΅ Π΄Π»Ρ ΠΎΠΏΡΠ΅Π΄Π΅Π»Π΅Π½ΠΈΡ ΡΠ»ΠΎΠ²Π΅ΡΠ½ΡΡ
Π°ΡΡΠΎΡΠΈΠ°ΡΠΈΠΉ, ΠΊΠΎΡΠΎΡΠΎΠ΅ ΡΠΌΠΎΠ³ΡΡ ΠΈΡΠΏΠΎΠ»ΡΠ·ΠΎΠ²Π°ΡΡ ΠΊΠΎΠ½Π΅ΡΠ½ΡΠ΅ ΠΏΠΎΠ»ΡΠ·ΠΎΠ²Π°ΡΠ΅Π»ΠΈ. Π£ΡΠΈΡΡΠ²Π°Ρ, ΡΡΠΎ ΡΡΡΡΠΊΠΈΠΉ ΠΈ Π°Π½Π³Π»ΠΈΠΉΡΠΊΠΈΠΉ ΡΠ²Π»ΡΡΡΡΡ ΠΎΠ΄Π½ΠΈΠΌΠΈ ΠΈΠ· Π½Π°ΠΈΠ±ΠΎΠ»Π΅Π΅ ΡΠ°ΡΠΏΡΠΎΡΡΡΠ°Π½Π΅Π½Π½ΡΡ
ΡΠ·ΡΠΊΠΎΠ² Π² ΠΌΠΈΡΠ΅, Π° Π ΠΎΡΡΠΈΡ ΡΠ²Π»ΡΠ΅ΡΡΡ ΠΏΠΎΠΏΡΠ»ΡΡΠ½ΡΠΌ ΠΌΠ΅ΡΡΠΎΠΌ Π΄Π»Ρ ΠΈΠ½ΠΎΡΡΡΠ°Π½Π½ΡΡ
ΡΡΡΠ΄Π΅Π½ΡΠΎΠ² ΡΠΎ Π²ΡΠ΅Π³ΠΎ ΠΌΠΈΡΠ°, ΡΡΠΎ ΠΏΠΎΡΠ»ΡΠΆΠΈΠ»ΠΎ Π·Π½Π°ΡΠΈΡΠ΅Π»ΡΠ½ΠΎΠΉ ΠΌΠΎΡΠΈΠ²Π°ΡΠΈΠ΅ΠΉ Π΄Π»Ρ ΡΠ°Π·ΡΠ°Π±ΠΎΡΠΊΠΈ ΠΈΠ½ΡΡΡΡΠΌΠ΅Π½ΡΠ° ΠΈΡΠΊΡΡΡΡΠ²Π΅Π½Π½ΠΎΠ³ΠΎ ΠΈΠ½ΡΠ΅Π»Π»Π΅ΠΊΡΠ°, ΠΊΠΎΡΠΎΡΡΠΉ ΠΏΠΎΠΌΠΎΠΆΠ΅Ρ Π»ΡΠ΄ΡΠΌ, ΠΈΠ·ΡΡΠ°ΡΡΠΈΠΌ ΡΡΡΡΠΊΠΈΠΉ ΡΠ·ΡΠΊ ΠΊΠ°ΠΊ Π°Π½Π³Π»ΠΎΠ³ΠΎΠ²ΠΎΡΡΡΠΈΠ΅, ΠΈΠ»ΠΈ ΠΈΠ·ΡΡΠ°ΡΡΠΈΠΌ Π°Π½Π³Π»ΠΈΠΉΡΠΊΠΈΠΉ ΡΠ·ΡΠΊ. ΠΊΠ°ΠΊ ΡΡΡΡΠΊΠΎΡΠ·ΡΡΠ½ΡΠ΅.Language learning is a remarkable endeavor that expands our horizons and allows us to connect with diverse cultures and people around the world. Traditionally, language education has relied on conventional methods such as textbooks, vocabulary drills, and language exchanges. However, with the advent of machine learning, a new era has dawned upon language instruction, offering innovative and efficient ways to accelerate language acquisition. One intriguing application of machine learning in language learning is the utilization of cognates, words that share similar meanings and spellings across different languages. To address this subject, this research paper proposes to facilitate the process of acquiring a second language with the help of artificial intelligence, particularly neural networks, which can identify and use words that are similar or identical in both the learner's first language and the target language. These words, known as lexical cognates which can facilitate language learning by providing a familiar point of reference for the learner and enabling them to associate new vocabulary with words they already know. By leveraging the power of neural networks to detect and utilize these cognates, learners will be able to accelerate their progress in acquiring a second language. Although the study of semantic similarity across different languages is not a new topic, our objective is to adopt a different approach for identifying Russian-English Lexical cognates and present the obtained results as a language learning tool, by using the lexical and semantic similarity data sample across languages to build a lexical cognates detection and words association model. Subsequently, depend on our analysis and results, will present a word association application that can be utilized by end users. Given that Russian and English are among the most widely spoken languages globally and that Russia is a popular destination for international students from around the world, it served as a significant motivation to develop an AI tool to assist individuals learning Russian as English speakers or learning English as Russian speakers
- β¦