Today’s interview features Mikhail Petrenko, Senior Data Architect at Adobe. Mikhail will be presenting Big Data Sampling: a Way to Make all of Your Data Useful Again.
Tell us about your presentation.
Mikhail: I feel efficient sampling on Apache Hadoop is underutilized by the Big Data community. I hope to give business, product and data specialists enough trust in the technology and enough ammunition to successfully argue that implementation of data sampling will benefit their enterprise while also providing sufficient information for actual implementation of the methodology.
What do you expect will be the single biggest takeaway of your presentation for attendees?
Mikhail: We can ingest more data, answer increasingly complex business questions and reduce costs all at the same time.
Tell us about your current role and how you interact with Hadoop.
Mikhail: As Senior Data Architect for Adobe Audience Manager, I view Hadoop as one of the core technologies of our solution. We use different aspects of the Hadoop ecosystem from plain HDFS operations to Hive, Pig and HBase. The Hadoop platform is intimately integrated with multiple other technologies both as data source and a target.
What are you most looking forward to at Hadoop Summit?
Mikhail: Conversations with vendors and sponsors, especially the ones I am not familiar with yet.
What other presentations are you most looking forward to attending?
Mikhail: Innovations in Hadoop MapReduce.