Fast Forward Labs research now available without a subscription
Moving forward, all new reports will be publicly available and free to download. In addition, we will be providing access to updated versions of older reports over time, so check back often to explore available free research.
Structural Time Series
Time series data is ubiquitous. This report examines generalized additive models, which give us a simple, flexible, and interpretable means for modeling time series by decomposing them into structural components. We look at the benefits and trade-offs of taking a curve-fitting approach to time series, and demonstrate its use via Facebook’s Prophet library on a demand forecasting problem.
In contrast to how humans learn, deep learning algorithms need vast amounts of data and compute and may yet struggle to generalize. Humans are successful in adapting quickly because they leverage their knowledge acquired from prior experience when faced with new problems. In this report, we explain how meta-learning can leverage previous knowledge acquired from data to solve novel tasks quickly and more efficiently during test time
Automated Question Answering
Automated question answering is a user-friendly way to extract information from data using natural language. Thanks to recent advances in natural language processing, question answering capabilities from unstructured text data have grown rapidly. This blog series offers a walk-through detailing the technical and practical aspects of building an end-to-end question answering system.
Causality for Machine Learning
The intersection of causal inference and machine learning is a rapidly expanding area of research that's already yielding capabilities to enable building more robust, reliable, and fair machine learning systems. This report offers an introduction to causal reasoning including causal graphs and invariant prediction and how to apply causal inference tools together with classic machine learning techniques in multiple use-cases.
Interpretability: 2020 Edition
Interpretability, or the ability to explain why and how a system makes a decision, can help us improve models, satisfy regulations, and build better products. Black-box techniques like deep learning have delivered breakthrough capabilities at the cost of interpretability. In this report, recently updated to include techniques like SHAP, we show how to make models interpretable without sacrificing their capabilities or accuracy.
Deep Learning for Anomaly Detection
From fraud detection to flagging abnormalities in imaging data, there are countless applications for automatic identification of abnormal data. This process can be challenging, especially when working with large, complex data. This report explores deep learning approaches (sequence models, VAEs, GANs) for anomaly detection, when to use them, performance benchmarks, and product possibilities.
Updated versions of older reports will be available for free in the future, so check back often.
Transfer Learning for NLP
Natural language processing (NLP) technologies can translate language, answer questions, and generate human-like text, but the underlying deep learning techniques require costly datasets, infrastructure, and expertise. In this report, we show how to use transfer learning to adapt existing models to any NLP application, making it easier to build high-performance NLP systems.
Deep Learning for Image Analysis - 2019 Edition
Convolutional neural networks (CNNs or ConvNets) excel at learning meaningful representations of features and concepts within images, making CNNs valuable for solving problems in multiple domains, from medical imaging to manufacturing. In this report, we show how to select the right deep learning models for image analysis tasks and techniques for debugging deep learning models.
In this report, we focus on federated learning, an approach for training machine learning models on distributed edge node data while ensuring privacy and minimizing communication costs.
In this report, we focus on multi-task learning, a new approach to machine learning that allows algorithms to master tasks in parallel.
In this report, we show how using the semantic content of items can help solve common recommendation pitfalls such as the cold start problem, and open up new product possibilities.
In this report, we show how to make models interpretable without sacrificing their capabilities or accuracy.
Here, we show how to use probabilistic programming and Bayesian inference to easily build tools that make better predictions for more effective decision making.
Learn how to use deep learning and embeddings to make text computable for a variety of business applications and products.
Deep learning: Image analysis
This report explores the history and current state of deep learning, explains how to apply it, and predicts future developments.
Probabilistic methods for realtime streams
Here, we explore probabilistic methods that offer highly efficient models for extracting value from streams of data as they are generated.
Natural language generation
In this report, we look at how machine systems can turn highly structured data into human language narrative.
Read the Fast Forward Labs blog
Keep up with tomorrow
Sign up for our monthly newsletter and get the latest on advances in applied artificial intelligence, as well as company news and events.