Abstract: Imagine a scenario where you are responsible for summarizing the information contained in tens of thousands of text documents. Perhaps you need to summarize product feedback, patent documents or technical specs and you are faced with the daunting task of sifting through an enormous pile of documents! This problem can be solved by using natural language processing and machine learning to programmatically extract text summaries.
In this talk, I will present an approach based on topic modeling. Specifically, we’ll use Latent Dirichlet Allocation (LDA) to generate short, human readable summaries spanning a range of discovered topics. We’ll describe the algorithm in the context of summarizing public feedback to pending government regulations, and demonstrate how this can be incorporated into an automated production pipeline. Along the way we’ll go over a brief history of text summarization, we’ll go over topic modelling and we'll discuss LDA, including other use cases and interesting visualizations.
Bio: Guilherme is a Data Scientist at Dataiku. He works out of the headquarters in NYC where he helps customers build and deploy predictive applications. Before joining Dataiku, he was a fellow at the Insight Data Science Fellowship program, and prior to that he worked in quantitative finance. He holds a PhD in applied mathematics.
Guilherme de Oliveira, PhD
Data Scientist at Dataiku