Motivation: Multiple sequence alignments (MSAs) of homologous sequences contain information on structural and functional constraints and their evolutionary histories. Despite their importance for many downstream tasks, such as structure prediction, MSA generation is often treated as a separate pre-processing step, without any guidance from the application it will be used for.
Results: Here, we implement a smooth and differentiable version of the Smith-Waterman pairwise alignment algorithm that enables jointly learning an MSA and a downstream machine learning system in an end-to-end fashion.
IEEE Trans Vis Comput Graph
January 2023
State-of-the-art neural language models can now be used to solve ad-hoc language tasks through zero-shot prompting without the need for supervised training. This approach has gained popularity in recent years, and researchers have demonstrated prompts that achieve strong accuracy on specific NLP tasks. However, finding a prompt for new tasks requires experimentation.
View Article and Find Full Text PDFIEEE Trans Vis Comput Graph
January 2022
Table2Text systems generate textual output based on structured data utilizing machine learning. These systems are essential for fluent natural language interfaces in tools such as virtual assistants; however, left to generate freely these ML systems often produce misleading or unexpected outputs. GenNI (Generation Negotiation Interface) is an interactive visual system for high-level human-AI collaboration in producing descriptive text.
View Article and Find Full Text PDFTwo-dimensional (2D) layered materials offer intriguing possibilities for novel physics and applications. Before any attempt at exploring the materials space in a systematic fashion, or combining insights from theory, computation, and experiment, a formal description of information about an assembly of arbitrary composition is required. Here, we introduce a domain-generic notation that is used to describe the space of 2D layered materials from monolayers to twisted assemblies of arbitrary composition, existent or not yet fabricated.
View Article and Find Full Text PDFIEEE Trans Vis Comput Graph
January 2020
Automation of tasks can have critical consequences when humans lose agency over decision processes. Deep learning models are particularly susceptible since current black-box approaches lack explainable reasoning. We argue that both the visual interface and model structure of deep learning systems need to take into account interaction design.
View Article and Find Full Text PDFIEEE Trans Vis Comput Graph
October 2018
Neural sequence-to-sequence models have proven to be accurate and robust for many sequence prediction tasks, and have become the standard approach for automatic translation of text. The models work with a five-stage blackbox pipeline that begins with encoding a source sequence to a vector space and then decoding out to a new target sequence. This process is now standard, but like many deep learning methods remains quite difficult to understand or debug.
View Article and Find Full Text PDFIEEE Trans Vis Comput Graph
January 2018
Recurrent neural networks, and in particular long short-term memory (LSTM) networks, are a remarkably effective tool for sequence modeling that learn a dense black-box hidden representation of their sequential input. Researchers interested in better understanding these models have studied the changes in hidden state representations over time and noticed some interpretable patterns but also significant noise. In this work, we present LSTMVis, a visual analysis tool for recurrent neural networks with a focus on understanding these hidden state dynamics.
View Article and Find Full Text PDF