View allAll Photos Tagged hadoop
FinData day
Alistair Croll (Solve For Interesting), Juan Huerta (Goldman Sachs Consumer Lending Group), Robert Passarella (Protege Partners), Giannina Segnini (Journalism School, Columbia University), Mar Cabra (International Consortium of Investigative Journalists), Anand Sanwal (CB Insights), Michael Casey (MIT Media Lab), Diane Chang (Intuit), Jeff McMillan (Morgan Stanley), Tanvi Singh (Credit Suisse), Kelley Yohe (Swift Capital), Michelle Bonat (Data Simply), Susan Woodward (Sand Hill Econometrics), Robert Passarella (Protege Partners)
What's next for Hadoop Panel at Structure Data Conference 2012 in NYC.
Photograph copyright Pınar Özger.
All rights reserved. Please contact via email to inquire about licensing for other usages.
Apache Solr on Hadoop is enabling organizations to collect, process and search larger, more varied data. Apache Spark is is making a large impact across the industry, changing the way we think about batch processing and replacing MapReduce in many cases. But how can production users easily migrate ingestion of HDFS data into Solr from MapReduce to Spark? How can they update and delete existing documents in Solr at scale? And how can they easily build flexible data ingestion pipelines?