8 research outputs found

    ChatMOF: An Autonomous AI System for Predicting and Generating Metal-Organic Frameworks

    Full text link
    ChatMOF is an autonomous Artificial Intelligence (AI) system that is built to predict and generate metal-organic frameworks (MOFs). By leveraging a large-scale language model (GPT-4 and GPT-3.5-turbo), ChatMOF extracts key details from textual inputs and delivers appropriate responses, thus eliminating the necessity for rigid structured queries. The system is comprised of three core components (i.e. an agent, a toolkit, and an evaluator) and it forms a robust pipeline that manages a variety of tasks, including data retrieval, property prediction, and structure generations. The study further explores the merits and constraints of using large language models (LLMs) AI system in material sciences using and showcases its transformative potential for future advancements

    Fine-tuned MOFTransformer model

    No full text
    Fine-tuned model (MOFTransformer) for ChatMOF</p

    PMTransformer: Universal Transfer Learning and Cross-material Few-shot Learning in Porous Materials

    No full text
    Porous materials have emerged as a promising solution for a wide range of energy and environmental applications. However, the asymmetric development in the field of MOFs has led to data imbalance when it comes to MOFs versus other porous materials such as COFs, PPNs, and zeolites. To address this issue, we introduce PMTransformer (Porous Material Transformer), a multi-modal pre-trained Transformer model pre-trained on a vast dataset of 1.9 million hypothetical porous materials, including metal-organic frameworks (MOFs), covalent-organic frameworks (COFs), porous polymer networks (PPNs), and zeolites. PMTransformer showcases remarkable transfer learning capabilities, resulting in state-of-the-art performance in predicting various porous material properties. To address the challenge of asymmetric data aggregation, we propose cross-material few-shot learning, which leverages the synergistic effect among different porous material classes to enhance fine-tuning performance with a limited number of examples. As a proof of concept, we demonstrate its effectiveness in predicting bandgap values of COFs using the available MOF data in the training set. Moreover, we established cross-material relationships in porous materials by predicting unseen properties of other classes of porous materials. Our approach presents a new pathway for understanding the underlying relationships between various classes of porous materials, paving the way toward a more comprehensive understanding and design of porous materials

    A Multi-modal Pre-training Transformer for Universal Transfer Learning in Metal-Organic Frameworks

    No full text
    Metal-organic frameworks (MOFs) are a class of crystalline porous materials that exhibit a vast chemical space due to their tunable molecular building blocks with diverse topologies. Given that an unlimited number of MOFs can, in principle, be synthesized, constructing structure-property relationships through a machine learning approach allows for efficient exploration of this vast chemical space, resulting in identifying optimal candidates with desired properties. In this work, we introduce MOFTransformer, a multi-model Transformer encoder pre-trained with 1 million hypothetical MOFs. This multi-modal model utilizes integrated atom-based graph and energy-grid embeddings to capture both local and global features of MOFs, respectively. By fine-tuning the pre-trained model with small datasets ranging from 5,000 to 20,000 MOFs, our model achieves state-of-the-art results for predicting across various properties including gas adsorption, diffusion, electronic properties, and even text-mined data. Beyond its universal transfer learning capabilities, MOFTransformer generates chemical insights by analyzing feature importance through attention scores within the self-attention layers. As such, this model can serve as a bedrock platform for other MOF researchers that seek to develop new machine learning models for their work

    Mining Insights on Metal-Organic Framework Synthesis from Scientific Literature Texts

    No full text
    Identifying optimal synthesis conditions for metal- organic frameworks (MOFs) is a major challenge that can serve as a bottleneck for new materials discovery and development. A trialand-error approach that relies on a chemist&apos;s intuition and knowledge has limitations in efficiency due to the large MOF synthesis space. To this end, 46,701 MOFs were data mined using our in-house developed code to extract their synthesis information from 28,565 MOF papers. The joint machine-learning/rule-based algorithm yields an average F1 score of 90.3% across different synthesis parameters (i.e., metal precursors, organic precursors, solvents, temperature, time, and composition). From this data set, a positive-unlabeled learning algorithm was developed to predict the synthesis of a given MOF material using synthesis conditions as inputs, and this algorithm successfully predicted successful synthesis in 83.1% of the synthesized data in the test set. Finally, our model correctly predicted three amorphous MOFs (with their representative experimental synthesis conditions) as having low synthesizability scores, while the counterpart crystalline MOFs showed high synthesizability scores. Our results show that big data extracted from the texts of MOF papers can be used to rationally predict synthesis conditions for these materials, which can accelerate the speed in which new MOFs are synthesized

    PMTransformer pre-trained model

    No full text
    Pretrained Machine learning models for MOFTransformer and PMTransformer.</p

    Hypothetical Porous Materials (COF, PPN, Zeolite)

    No full text
    Hypothetical porous material structures (COF, PPN, Zeolite) Construct using PORMAKE (https://github.com/Sangwon91/PORMAKE/tree/master/pormake) COF : 0.5M PPN : 0.3M Zeolite : 0.1M</p

    A Study on the Channel Expansion VAE for Content-Based Image Retrieval

    No full text
    Content-based image retrieval (CBIR) focuses on video searching with fine-tuning of pre-trained off-the-shelf features. CBIR is an intuitive method for image retrieval, although it still requires labeled datasets for fine-tuning due to the inefficiency caused by annotation. Therefore, we explored an unsupervised model for feature extraction of image contents. We used a variational auto-encoder (VAE) expanding channel of neural networks and studied the activation of layer outputs. In this study, the channel expansion method boosted the capability of image retrieval by exploring more kernels and selecting a layer of comparatively activated object region. The experiment included a comparison of channel expansion and visualization of each layer in the encoder network. The proposed model achieved (52.7%) mAP, which outperformed (36.5%) the existing VAE on the MNIST dataset
    corecore