Abstract
Molecular representation learning plays an essential role in cheminformatics. Recently, language model-based approaches have gained popularity as an alternative to traditional expert-designed features to encode molecules. However, these approaches only utilize a single molecular language for representation learning. Motivated by the fact that a given molecule can be described using different languages such as Simplified Molecular Line Entry System (SMILES), the International Union of Pure and Applied Chemistry (IUPAC), and the IUPAC International Chemical Identifier (InChI), we propose a multilingual molecular embedding generation approach called MM-Deacon (multilingual molecular domain embedding analysis via contrastive learning). MM-Deacon is pre-trained using SMILES and IUPAC as two different languages on large-scale molecules. We evaluated the robustness of our method on seven molecular property prediction tasks from MoleculeNet benchmark, zero-shot cross-lingual retrieval, and a drug-drug interaction prediction task.
Cite
CITATION STYLE
Guo, Z., Sharma, P., Martinez, A., Du, L., & Abraham, R. (2022). Multilingual Molecular Representation Learning via Contrastive Pre-training. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 3441–3453). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.242
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.