Amotz Maimon is Yahoo!’s Chief Architect, and leads Yahoo!’s Tech Council. Prior to Yahoo!, he was responsible for Strategy & Emerging Technologies in the xCP business unit at EMC’s Intelligent Information Group, where he developed the next generation cloud offerings and initiated the active and growing xCP developer community. Before joining the xCP business unit, Amotz served as GM of EMC IIG’s Business Process Management group. Before joining EMC, Amotz was CTO and SVP of Products & Engineering at Radiance, where he managed all of the company’s product and engineering operations, including development, R&D, product management, and customer relationships. Earlier in his career, Amotz was involved in a start-up acquired by Ventro, and held various technical and management positions at Applied Materials.
Amotz holds graduate degrees in Mathematics and Computer Science from the University of Tel-Aviv, Israel and an MBA from Pepperdine University’s Grazadio School of Business.
Arun is a Apache Hadoop PMC member and has been a full time contributor to the project since the inception in 2006. He is also the lead of the MapReduce project and has focused on building NextGen MapReduce (YARN). Prior to co-founding Hortonworks, Arun was responsible for all MapReduce code and configuration deployed across the 42,000+ servers at Yahoo!. In essence, he was responsible for running Apache Hadoop’s MapReduce as a service for Yahoo!. Also, he jointly holds the current world sortingrecord using Apache Hadoop. Follow Arun on Twitter: @acmurthy.
Herb Cunitz brings 25 years of experience building sales, consulting, training, marketing and operations teams for emerging enterprise software companies. Herb is responsible for global sales, marketing and support activities at Hortonworks. Herb was most recently responsible for global field operations for the vFabric Cloud Application Platform division of VMware and was responsible for global sales of the application run-time and data products/services that as part of the Cloud Application Platform. Prior to that Herb ran worldwide field operations for SpringSource – an open source software company acquired by VMware in 2009. Herb also spent 10 years at IBM in NY in various positions, including Sales/Marketing Manager for NY and US Marketing Executive for Desktop Operating Systems. Herb has an MBA in Finance from New York University Stern School of Business and a BS in Electrical Engineering from Columbia University in New York.
As Wikibon’s lead Big Data analyst, Jeff Kelly applies a critical eye to trends and developments in the Big Data and business analytics markets, with a strong focus on helping practitioners deliver business value. Jeff’s research includes market analysis, emerging technologies, enterprise Big Data case studies, and more. He also appears frequently on theCUBE to share his insights. Prior to joining Wikibon, Jeff spent seven years as a writer and editor at TechTarget, where he covered a number of business and IT topics including IT services, mobile computing, data management and business intelligence. He holds a BA from Providence College and an MA from Northeastern University.
John Schitka, is a Solution Marketing Manager on the SAP Big Data Solution Marketing team. His focus in the SAP Big Data arena is largely on Hadoop and SAP HANA smart data access capabilities. A graduate of McMaster University, he holds an MBA from the University of Windsor. He has worked in product marketing and product management in the high tech arena for a number of years, taught at a private college and has co-authored a number of published text books. He has a true love of technology and all that it has to offer the world.
Merv Adrian is an Analyst following database, big data, NoSQL and adjacent technologies as extreme information processing transforms assumptions about how to activate unused information assets. He also monitors the way the software/hardware boundary for information platforms is shifting as more processing moves into the hardware layer and appliances change purchasing patterns. Prior to Gartner, Mr. Adrian headed IT Market Strategy, which he founded to provide professional services to information technology strategy and marketing professionals. Earlier, he ran technology research for Forrester, following its acquisition of Giga Information Group. He spent a decade in the software industry after several years as a programmer on Wall Street.
Oliver Ratzesberger leads the software teams for Teradata Labs, including the Teradata Database, Aster, Client tools, Viewpoint as well as Hadoop integration.
Before joining Teradata, he was at Sears Holdings, where he drove a large analytics effort to consolidate systems into a newly redesigned Unified Data Architecture. Prior to Sears Holdings, he spent seven years at eBay, where he was responsible for its data warehouse and big data platforms. During his tenure at eBay, he led eBay’s expansion of analytics and was responsible for the co-development of the Extreme Data Appliance as part of eBay’s Singularity project. He also led the Hadoop platform engineering teams and drove the initial integration for Teradata and Hadoop at eBay.
As corporate vice president of program management for Microsoft Data Platform, Quentin Clark oversees the design and delivery of the entire family of SQL Server products. Leading a team of technical engineers, his responsibilities include product direction and definition through program management, user experience design, and customer engagement programs. This spans SQL Server’s work in all workloads – databases, integration and business intelligence, as well as the release forms of the product – software, appliance and the cloud. Clark joined Microsoft in 1994, and held a variety of roles mostly in systems technology prior to joining SQL Server, including the Internet Information Server team, Application Center and System Center. In 2002, he joined SQL Server to lead work to bring unstructured data capabilities into the data platform, a vision realized over the couple releases of SQL Server including SQL Server 2012. In 2006 he took the position of General Manager overseeing the core relational database engine, where he lead the SQL Server 2008 and 2012 releases, and drove the business into key new areas such as the appliance market, scale-out data warehousing and continuous event processing. Clark graduated with a degree in Natural Sciences (Physics) from the University of Massachusetts in Amherst and doubled-majored in Computer Science. He lives in Bellevue, WA with his wife and two boys.
Rob is the former chief operating officer of both SpringSource and JBoss, two of the most commercially successful open source companies in history. Rob also served in senior roles at Oracle, where he directed a $1 billion sales organization, I2 and Manhattan Associates. He currently sits on the boards of Black Duck and EnterpriseDB, two successful and well-regarded vendors in the open source industry. Rob has a Bachelor’s degree in Business Management from Jacksonville State University and currently holds the title of Entrepreneur-in-Residence at the University of Georgia’s Terry College of Business.
Shaun has more than 20 years of experience in the software industry, with a track record of building early stage and midsize software companies into successful market leaders. Shaun was previously VP of Product Strategy at VMware where he focused on defining VMware’s cloud application platform vision. Shaun has also held VP and Director level positions at SpringSource, Red Hat, JBoss, Princeton Softech, HP, Bluestone Software, and Primavera Systems. He holds a B.S. in Electrical Engineering from Drexel University. Follow Shaun on Twitter: @shaunconnolly.
Tom Davenport is the President’s Distinguished Professor of Information Technology and Management at Babson College, the co-founder of the International Institute for Analytics, a Fellow of the MIT Center for Digital Business, and a Senior Advisor to Deloitte Analytics. He teaches on analytics and big data for executive programs at Babson, Harvard Business School, MIT Sloan School, and Boston University. He was among the first to write about both business process reengineering and knowledge management. More recently, he pioneered the concept of “competing on analytics” with his best-selling 2006 Harvard Business Review article (and his 2007 book by the same name). His most recent book is Big Data@Work, from Harvard Business Review Press. He wrote or edited sixteen other books and over 100 articles for Harvard Business Review, Sloan Management Review, the Financial Times, and many other publications. In 2003 he was named one of the world’s “Top 25 Consultants” by Consulting magazine. In 2005 Optimize magazine’s readers named him among the top 3 business/technology analysts in the world. In 2007 and 2008 he was named one of the 100 most influential people in the IT industry by Ziff-Davis magazines. In 2012 he was named one of the world’s top fifty business school professors by Fortune magazine.
Tim is Senior Vice President, Infrastructure Business Group at Red Hat (NYSE: RHT), the world’s leading provider of open source solutions. He is responsible for Red Hat’s infrastructure and Infrastructure-as-a-Service (IaaS) businesses, including Linux Platform, Virtualization and OpenStack, and Cloud Management product lines. Prior to rejoining Red Hat, Tim was CEO of Black Duck Software, the leader in automating management, governance and the secure use of open source software. He has over 30 years of technology leadership experience including several years previously at Red Hat as CMO, where he was instrumental in expanding into the developer and middleware markets. Tim has also held executive positions at EqualLogic/Dell, Avaki and Macromedia, and spent the early part of his career at Compaq and Digital Equipment Corporation. Tim also sits on the Board of Monotype (Nasdaq: TYPE), Actuate (Nasdaq: BIRT) and on the Roger Williams University Board of Trustees. In 2009 Tim was named one of the Most Influential People in the Open Source Software Industry, in a survey by Mindtouch. In 2010 he was named a COPU Think Tank Advisor, which recognizes the key global thought leaders in open source software. Also in 2010, the Mass. Technology Leadership Council named him a finalist for CEO of the Year. Tim holds an M.B.A. from Babson College, and a B.S., summa cum laude, from Roger Williams University.
Victor Nilson is Senior Vice President for Big Data at AT&T, leading a centralized organization dedicated to advancing AT&T’s Big Data capabilities. His diverse business experience ranges from the software and travel industries to several roles at AT&T. He has served in leadership at AT&T and formerly Cingular in information technology, architecture, engineering, common services integration and application development. Nilson holds a Master’s of Business Administration from Southern Methodist University. He is responsible for a number of patents in technology and business.
Aaron Davidson is an Apache Spark committer and software engineer at Databricks. He has implemented Spark standalone cluster fault tolerance and shuffle file consolidation, and has helped in the design, implementation, and testing of Spark`s external sorting and driver fault tolerance.
Aaron T. Myers is a Software Engineer at Cloudera and an Apache Hadoop Committer. Aaron’s work is primarily focused on HDFS. Prior to joining Cloudera, Aaron was a Software Engineer and VP of Engineering at Amie Street, where he worked on all components of the software stack, including operations, infrastructure, and customer-facing feature development. Aaron holds both an Sc.B. and Sc.M. in Computer Science from Brown University. http://blog.cloudera.com/blog/2012/08/meet-the-engineer-aaron-t-myers/
Abhijit is the Chief Architect of Big Data Systems at Rocket Fuel that enabled learning and gaining insight over several petabytes of data everyday. Prior to Rocket Fuel, he was an architect working on Yahoo?s Behavioral Targeting Platform that builds machine-learning models over big data. He holds Ph.D. in Computer Science specialized in databases and Approximate Query Processing. Abhijit is a co-author of ?decision support systems? which is used as a textbook in many universities. He is author of 10+ research papers published in prestigious international database conferences. Abhijit also co-authored a paper that won of SIGMOD-2005 best paper award.
Adam is a Deep Learning specialist based in San Francisco assisting Fortune 500 companies, hedge funds, PR firms and startup accelerators with their machine learning projects. Adam has a strong track record helping companies handle and interpret big real-time data. Adam has been a computer nerd since he was 13 and actively contributes to the open source community.
Adam Kawa works as Data Engineer at Spotify, where his main responsibility is to maintain one of the largest Hadoop-YARN clusters in Europe. Every so often, he implements and troubleshoots Python MapReduce, Hive and Pig jobs. Adam is a frequent speaker at Hadoop conferences and Hadoop User Groups meetups. He co-organizes Stockholm and Warsaw Hadoop User Groups. He regularly blogs about the Hadoop ecosystem at HakunaMapData.com.
Adriana Zubiri is a Program Director in the Information Management Big Data group at the IBM Toronto Software Lab. In her current role, Adriana is responsible for driving the software development execution for Big SQL, IBM’s SQL-on-Hadoop solution, which is as part of the IBM BigInsights offering. Adriana is recognized within the industry as an expert in the area of data warehouse performance, based on her extensive work with clients, her numerous papers and patents on database technology. Adriana received a MSc in Computer Science from University of Alberta as well as an honors degree in Computer Science from Universidad CAECE in Buenos Aires, Argentina.
Alan works in Pythian’s CTO office, helping clients architect their data infrastructure. His focus is on assessing, selecting and integrating cutting-edge technologies including Hadoop and ecosystem projects. When he isn’t working on client systems, Alan develops web applications on a variety of platforms to solve Pythian’s and his own problems.
Alan is a co-founder at Hortonworks and an original member of the engineering team that took Pig from a Yahoo! Labs research project to a successful Apache open source project. Alan also designed HCatalog and guided its adoption as an Apache Incubator project. Alan has a BS in Mathematics from Oregon State University and a MA in Theology from Fuller Theological Seminary. He is also the author of Programming Pig, a book from O’Reilly Press.
Alex Moundalexis is a Solutions Architect for Cloudera Government Solutions and has spent the last year installing and configuring Hadoop clusters across the country for a variety of commercial and government customers. Before entering the land of Big Data, Alex spent the better part of ten years wrangling Linux server farms and writing Perl as a contractor to the Department of Defense and Department of Justice. He likes shiny objects.
Alexis Roos is a Senior Solutions Architect focusing on Big Data solutions at Concurrent, Inc. He has over 18 years of experience in software and sales engineering, helping both Fortune 500 firms and start-ups build new products that leverage Big Data, application infrastructure, security, databases and mobile technologies. Prior, Alexis worked for Sun Microsystems and Oracle for over 13 years, and has also spent time at Couchbase and several large systems integrators over in Europe. Alexis has spoken at dozens of conferences as well as university courses and holds a Master’s Degree in Computer Science with a Cognitive Science emphasis.
Allen has been working on scientific computing and BigData problems for 15 years, first with Beowulf clusters and Hadoop since 2008. Allen created one of the first video discovery platforms for mobile phones, and also built the largest public database of human gene expression data. As an open source developer, Allen has authored R and Perl libraries and most recently contributed to Storm. Allen has founded and co-founded multiple machine learning and biotech companies, and he holds a Ph.D. in Human Genetics from the School of Medicine at UCLA.
Allen Wittenauer has been involved with Apache Hadoop since May 2007, when he was hired by Yahoo! to bring large-scale operational experience to the fledgling project. His work there helped create the basic blueprints that almost all Hadoop deployments follow today. At LinkedIn, his experience provided key insight and a foundation to its award-winning data science team.
Amareshwari is currently working as Architect in platform team at Inmobi, where she works on Hadoop and related projects for data collection and analytics. She is member of Apache Hadoop PMC and is Apache Hive committer. She has been working on Hadoop and its eco system since 2007. Prior to Inmobi, she was working with Yahoo! in core Hadoop team. She holds bachelor`s degree in computer science and engineering from National institute of technology, Waragal, India; and master`s degree in Internet science and engineering from Indian Institute of Science (IISc), Bangalore, India.
Amritashwar Lal is a Product Manager at Yahoo where he is engaged in building a high class and robust Hadoop infrastructure services. He has eight years of experience across HSBC, Oracle and Google in developing products and platforms for high growth enterprises. He earned his MBA from Carnegie Mellon University Tepper School of Business in Marketing and Entrepreneurship.
Andrew Nelson currently works for VMware as an SE specialist focused on enterprise and strategic accounts and is a VMware Certified Design Expert (VCDX #33) and CISSP. With a deep background in storage, networking, security and availability, he has over 18 years of IT experience and continues to work on virtualizing next-gen applications in the fields of High-Performance Computing and Big Data. Most recently he presented at VMworld US and EMEA around virtualized HPC with UCSF and he is a regular speaker at VMworld, EMCworld, and VMUGs across the western US. Prior to joining VMware, he worked as the virtualization architect for the US Marine Corps and has focused his time working on accounts across the US Department of Defense and private sectors.
Andy Feng is a Distinguished Architect at Yahoo!, and a Committer at Apache Storm project. He is leading architecture design of Yahoo big-data platform.
Ankur Gupta is an IT Director at Sears and Global Head of Sales, Marketing and Operations at MetaScale, a Big Data technology Subsidiary of Sears Holdings. Ankur is leading efforts to accelerate Big Data efforts for other enterprises, leveraging learning from implementing Hadoop and other Big Data technologies at Sears. Before moving into this role, Ankur has led several other major monetization initiatives at Sears in various businesses. Prior to Sears, Ankur worked with IBM Global Services in India and US. Ankur received an MBA from Duke University and a degree in Mechanical Engineering from Indian Institute of Technology, Roorkee.
Ari Zilka, CTO of Hortonworks, has more than 20 years of software development expertise and a deep understanding of open source, enterprise software, and the execution required to build successful products. Ari was previously founder and CTO at Terracotta. Previously, Ari was an Entrepreneur-in-Residence at Accel Partners.
Avery has a PhD from Northwestern University in the area of parallel computing. He worked at Yahoo! Search for four years on the web map analytics platform, large-scale ad hoc serving infrastructure, and cluster management. During the past two and a half years, he has been working at Facebook in the general area of big data computational frameworks (Corona – MapReduce, Giraph).
Bernardo de Seabra heads the data team at BrightRoll, the world’s largest video advertising platform. As a Staff Software Engineer, he is faced with the challenge of extracting meaning from BrightRoll’s ever increasing amounts of data (X TB per month). Prior to BrightRoll, Bernardo worked for Playdom, where he was the first engineer to work on the “Next Generation” mobile services platform project, a project aimed to provide a back-end platform for games to plug into and share business logic and infrastructure. Bernardo has a BA in Computer Science from the University of California Berkeley.
Bikas has been working in the Apache Hadoop ecosystem since 2011 and is a committer/PMC member of the Apache Hadoop and Tez projects. He has been a key contributor in making Hadoop run natively on Windows and has focused on YARN and the Hadoop compute stack, with special interest in Tez. Prior to Hadoop, he has worked extensively on the Dryad distributed data processing framework that runs on some of the world’s largest clusters as part of Microsoft Bing infrastructure. @bikassaha
Bill Yetman has served as VP of Engineering at Ancestry.com since January 2014. Bill has held multiple positions with Ancestry.com from August 2002, including Senior Director of Engineering, Director of Sites, Mobile and APIs, Director of Ad Operations and Ad Sales, Senior Software Manager of eCommerce and Senior Software Developer. Prior to joining Ancestry.com, he held several developer and programmer roles with Coresoft Technologies, Inc., Novell/Word Perfect, Fujitsu Systems of America and NCR. Bill holds a B.S. in Computer Science and a B.A. in Psychology from San Diego State University.
Bobby Evans is a Principal Engineer at Yahoo! and a Hadoop PMC member at the Apache Software Foundation.
Brian is a Principal Solutions Architect with SAS, focusing on Big Data and Hadoop. With more than 20 years in analytics and IT architecture, Brian has worked on big data projects around the world, including the 2010 U.S. Census. Prior to joining SAS, Brian was the Chief Technical Officer of Mustard Tree Instruments, developing an analytical and computing platform for the pharmaceutical and chemical industries. Brian holds a B.S. in Statistics from NC State and a M.S. in Information Systems from The George Washington University.
I specialize in writing software and solving problems where there are either scalability concerns due to large amounts of traffic or large amounts of data. I have a particular passion for data science problems or any thing vaguely mathematical. As a Principal Architect focused on data science, I spend time with a variety of clients, large and small, mentoring and helping them use Hadoop to solve their problems. I have specialized in the past in Oil & Gas and Healthcare.
Chintan holds Masters of Technology from IIIT. A Passionate big data enthusiast, with 8 years of experience, works with Syntel. He works on Research & development. He is also involved in building the Industry solutions
Chris Mutchler is a seventeen-year veteran of utilizing Open Source virtualization technologies in large enterprise environments. Chris currently works for Adobe Systems where he is a Compute Platform Engineer, specializing in building out turnkey Infrastructure-as-a-Service and Platform-as-a-Service offerings. Before joining Adobe, he managed a team of system administrators that grew a virtualized web-hosting business from a few hundred servers to over ten thousand. He has received numerous awards for his contributions to his profession. An important aspect of his life is family – his wife, three daughters, and three sons.
Chris is a software engineer at Hortonworks and an Apache Hadoop committer. He is an active contributor to HDFS, YARN, and MapReduce. Prior to Hortonworks, Chris worked for Disney, where he deployed Hadoop, developed data management solutions on top of it, and was responsible for operational support.
Clay comes from a diverse background in systems infrastructure and analytics. His team built out an automated bare-metal Solaris deployment tool for Solaris engineering labs and his contributions were core to the OpenSolaris Automated Installer at Sun Microsystems. Providing a good introduction to Hadoop, his team built out a financial portfolio analytics product at Opera Solution on top of Oozie, Pig, Python streaming and Sqoop. Merging the two, his team is now building out infrastructure for low-latency HBase and high ingest data warehousing using much of the Hadoop ecosystem at Bloomberg.
Costin Leau is an engineer at Elasticsearch, currently working with NoSQL and Big Data technologies. An open-source veteran, Costin led various Spring projects (Spring OSGi, GemFire, Redis, Hadoop) and authored an OSGi spec. Speaker at various editions of EclipseCon/OSGi DevCon, JavaOne, Devoxx/Javapolis, JavaZone, SpringOne, TSSJS on Java/Hadoop/Spring related topics.
Dan McClary currently serves as Principal Product Manager for Big Data and Hadoop at Oracle. Prior to joining Oracle he served as Director of Business Intelligence at Red Robot Labs in Palo Alto, CA. He previously was as a Visiting Scholar at Northwestern University and the Howard Hughes Medical Institute where his research in Complex Systems focused on applying Hadoop to large scale graph problems. Dr. McClary received his PhD in Computer Science from Arizona State University, where his work centered on adaptive optimization of mobile and ad hoc networks. He holds an M.S. in Computer Science from Arizona State focused on hard real-time schedulability in distributed systems and was founder of imgSurf, a biometrics and electronic medical record company. Dr. McClary contributes to open source projects such as Tinkerpop Blueprints and Apache Spark.
Daniel is an Apache Pig PMC member/committer involved with Pig for 5 years at Yahoo and now at Hortonworks. He has a PhD in Computer Science with specialization in computer security, data mining and distributed computing from University of Central Florida. He is interested in data science, large scale processing, Hadoop, Pig, Hive, and more.
Daniel works in the Cloudera training team building Cloudera’s developer and data science Cloudera Certified Professional certifications. Daniel also has a long history as a software engineer in the high performance computing space and has been kicking around big data since about 2009. Prior to Cloudera, Daniel spent more than a decade at Sun doing various engineering and product management roles and speaking at conferences. Daniel has a BE in EE/CS from Vanderbilt and is just finishing an MSCS from Stanford.
Danil has been working with databases and information systems since his early years in university, where he received a Masters Degree in Applied Math. Danil has 7 years of experience architecting, building and supporting large mission-critical data platforms using various flavours of MySQL, Hadoop and MongoDB. He is also the author of “Hadoop Cluster Deployment” book. Besides databases Danil is interested in functional programming, machine learning and rock climbing.
David Chaiken comes to Altiscale from Yahoo!, where he served as Chief Architect. At Yahoo, he led teams building consumer advertising and media systems with Apache Hadoop at their core. Over his career, David has also built voice-search products for consumers, mobile applications for enterprises, network management systems, project management software, large-scale multiprocessor architectures, a tablet computer, and several other information appliances. David earned a BS in Mathematics from Brown University and a PhD in Electrical Engineering and Computer Science from MIT.
David is a software engineer on the Hadoop Development team in LinkedIn`s Data Analytics Infrastructure group. He has previously worked at Microsoft in the Windows Kernel Platform Group.
David P. Mariani is CEO of AtScale, Inc., an incubating software startup focused on bringing business intelligence into age of Hadoop. Prior to AtScale, David was VP Engineering of Klout, a social analytics data service that scores over 450 million user profiles daily and collects over 12 billion events across the social web. Previously, David ran the analytics and data pipelines for Yahoo!`s consumer sites and advertising services, where he built the world`s largest cube and drove early Hadoop development and adoption.
Enis Soztutar is an Apache HBase, Hadoop, and Gora PMC member and member of the Apache Software Foundation. He has been using and developing Hadoop ecosystem projects since 2007. He is currently working at Hortonworks as a part of the HBase engineering team.
Garrett joined WibiData in 2010 and focuses on distributed infrastructure and algorithms. Previously he worked at Google in Mountain View, CA and New York City where he was the tech lead of the personalized recommendations team. Garrett’s areas of interest include natural language processing, machine learning and data mining.
A mathematician and seasoned technology executive, George Corugedo has over 20 years of business and technical expertise. As co-founder and CTO of RedPoint Global, George leads the development of RedPoint’s Data Management and Data Management for Hadoop applications, and its Convergent Marketing Platform™. A former math professor, George left academia to co-found Accenture’s Customer Insight Practice, which specialized in strategic data utilization, analytics and customer strategy. Previous positions included director of client delivery at ClarityBlue, a provider of hosted customer intelligence solutions to commercial enterprises, and COO/CIO of Riscuity, a receivables management company specializing in utilizing analytics to drive collections.
Gera graduated from Universitt des Saarlandes with PhD in Computer Science. He has published various papers related to Database and transactional systems. He has worked at Oracle, MapR and now at Twitter.
Gopal Vijayaraghavan is a late entry into the hadoop game, having started working on it in 2012. He works on hive and Tez as part of the Stinger initiative, fixing query performance at scale.
Govind Kamat is a member of the Performance Engineering Team at Cloudera, focusing on Hadoop and HBase performance and scalability. His experience includes the development of large-scale software systems, microprocessor architecture, compilers and electronic design. Before Cloudera, he was a member of the Performance Engineering Group at Yahoo!
Gunther Hagleitner has been contributing to various hadoop projects for over four years both at Yahoo! as well as Hortonworks. He is an active committer in the Apache Hive project as well as a PMC member of the Apache Tez project. Before Hadoop, Gunther has been working on database technology for more than a decade. At Hortonworks he is leading Hive efforts in the Stinger project – delivering performance and SQL capabilities in the ecosystem. Gunther holds has a MS in Mathematics from the University of Konstanz.
Hitesh Shah currently works on various things related to Apache Hadoop at Hortonworks with his primary focus on Apache Tez and Apache Hadoop YARN. He is a PMC member and committer for the Apache Hadoop, Tez and Ambari projects. Earlier to that, he spent close to a decade at Yahoo! building various frameworks all the way from data storage platforms for social content to a multi-threaded event-driven framework for building high throughput advertising serving platforms.
Hyunsik Choi, Ph.D., is an Apache Tajo committer and PPMC member. He is the Director of Research at Gruter Corp., a big data company based in Seoul, South Korea, and has made a significant contribution to the Tajo query plan optimizer, vectorized query engine, and new-generation hardware integration features. His recent Tajo contributions focus on runtime query compilation techniques using LLVM, as well as SSD and SIMD hardware capabilities.
Ingo Mierswa is an industry-veteran data scientist since he began to develop RapidMiner predictive analytics software at the Artificial Intelligence Division of the University of Dortmund, Germany. Mierswa, the scientist, has authored numerous award-winning publications about predictive analytics and Big Data. Mierswa, the entrepreneur, is the founder of RapidMiner. Under his leadership, the company has grown up to 300 percent per year over the past five years. In 2012, he spearheaded the go-international strategy with the opening of offices in the U.S.
Jagane Sundar is an expert in Big Data, Cloud, Virtualization, and Networking technologies. He co-founded AltoStor, a Hadoop based Big Data Storage company acquired by WANdisco. His Hadoop experience began as Director of Hadoop Performance and Operability at Yahoo! Jagane has such accomplishments to his credit as the creation of Livebackup, an open source KVM VMbackup solution, a user mode TCP Stack for Precision I/O, NFS and PPP clients and parts of the TCP stack for Sun MicroSystems’ JavaOS and the VxDTCP, a 32bit VxD based TCP Stack for Windows. Jagane received his B.E. in ECE from the College of Engineering, Guindy, Anna University, Chennai.
James Sirota loves to code, tinker, and experiment. In his spare time he works as a Big Data Solutions Architect and a Data Scientist at Cisco. He is a lead engineer on the big data platform for OpenSOC, a new security analytics platform from Cisco. James has over 10 years of experience as a software developer and a security engineer. He has an M.S. in Systems Engineering from the University of Southern California and a B.S. in Computer Science from Arizona State University.
James Taylor is an engineer at Salesforce.com in the Big Data Group. He founded the Apache Phoenix project and has lead the development effort on that for the past several years. Prior to working at Salesforce.com, James worked at BEA Systems on federated query processing systems and event driven programming models. He lives in San Francisco with his wife and two daughters.
Janos is CTO at SequenceIQ – a young startup with the mission statement of simplifying provisioning, development and build QoS on Hadoop. Before co-founding SequenceIQ he was a Solutions Architect at EPAM Systems. He is an open source advocate and contributor, a Hadoop YARN evangelist and a keen surfer and freeskier. He holds a Master’s Degree in Computer Science, specialized on distributed systems.
Jason is an Apache Hadoop PMC member and has been contributing to Hadoop for the last 2 years. He is one of the lead Hadoop developers at Yahoo with a primary focus of running YARN and MapReduce on large-scale clusters.
The speaker Jay Tang is currently Director of Big Data Engineering at PayPal leading its big data effort. He is passionate about data. Jay was a member of the original Hadoop team at Yahoo! building and managing the world`s largest Hadoop cluster at that time. He has built large parallel databases at IBM/Informix/Yahoo! and BI products at Oracle/Hyperion.
Jeff Graham is a Senior Advisor of Data Analytics who is responsible for Big Data architecture at Cardinal Health. He has over 20 years experience in BI, database performance tuning as well as systems and application development.
Jeffrey Heer is an Associate Professor of Computer Science at the University of Washington, where he works on human-computer interaction, visualization, and social computing. The visualization tools developed by his lab (D3.js, Protovis, Flare, Prefuse) are used by researchers, companies, and thousands of data enthusiasts around the world. His group’s papers have received awards at the premier venues in human-computer interaction and information visualization (ACM CHI, ACM UIST, IEEE InfoVis, IEEE VAST, EuroVis). In 2009, Jeff was named to MIT Technology Review’s TR35, and in 2012 he was named a Sloan Foundation Research Fellow. He holds BS, MS, and PhD degrees in computer science from the University of California, Berkeley. Jeff is also the Co-Founder and Chief Experience Officer at Trifacta.
Jennifer Lim has over 14 years of experience in large scale enterprise data warehousing and analytics. Most recently, she was a Research Scientist for the Sprint Advanced Analytics Lab and is now acting as a Lead Technology Architect, focusing on upgrading the analytics infrastructure in support of all those great use cases being discovered in the research lab.
Jian He is a committer to the Apache Hadoop project. He is a Software Engineer at Hortonworks in the MapReduce team and mostly focuses on YARN development. Prior to joining Hortonworks, he received a Masters degree in Computer Science from Brown University.
Jim Vogt has more than 25 years of leadership experience in both start-up and established corporations. Most recently, Jim served as senior vice president and general manager of the Cloud Services business unit at Blue Coat Systems. Prior to Blue Coat, he served as president and CEO at Trapeze Networks, which was acquired by Belden, Inc. He was also president and CEO at data encryption start-up Ingrian Networks (acquired in April, 2008 by SafeNet). Jim holds a BS degree in electrical engineering from the University of Nevada, and an MBA from Santa Clara University.
Jitendra Pandey works at Hortonworks Inc and has been contributing to Hadoop for around 5 years. His current area of focus is Hive performance improvements. Jitendra is a committer and PMC member for Apache Hadoop. He is also a committer and PMC member for Apache Ambari. Jitendra`s contributions include vectorized query processing in Hive, Ambari development, Hadoop security infrastructure, federated HDFS, wire compatibility and high availability of HDFS. Prior to Hortonworks, Jitendra worked at Yahoo on Big Data infrastructure and applications
Jiwon Seo is a CS PhD student in Stanford university. He is interested in distributed systems, large-scale data mining, and data visualization. With his advisor, professor Monica Lam, he designed and implemented distributed query language, called SociaLite. With its high-level semantic, SociaLite makes it easy to implement efficient code for large-scale data processing. He is interested in applying SociaLite for large-scale graph analysis and data mining. Jiwon has a Bachelor`s degree in Electrical Engineering from Seoul National University, and a Master`s degree in Electrical Engineering from Stanford University.
Joe Rao has been working with Teradata since the beginning of its efforts to bring enterprise-friendly Hadoop to the market. Working in presales, post sales, and support roles across a wide range of large clients, Joe has experienced the challenges facing enterprise adoption of Hadoop. Before his work at Teradata, Joe studied Computer Science at the Georgia Institute of Technology.
Joey Echeverria is Cloudera`s Chief Architect for Public Sector where he coordinates with Cloudera`s Customers and Partners as well as Cloudera`s Product, Engineering, and Field teams to speed up the time it takes to move Hadoop applications to production. Previously Joey was a Principal Solutions Architect where he worked directly with customers to deploy production Hadoop clusters and solve a diverse range of business and technical problems. Joey joined Cloudera from the NSA where he worked on data mining, network security, and clustered data processing using Hadoop.
With over 15 years in advanced analytical applications and architecture, John is dedicated to helping organizations become more data-driven. He combines deep expertise in analytics and data science with business acumen and dynamic engineering leadership.
John is a senior engineer at Hortonworks and a member of the Savanna development team. John is also a committer on the Apache Ambari project where he designed the current Ambari REST API. He has 15 years of experience developing commercial middleware systems with a focus on distributed transaction processing.
John Williams leads the Platform Operations team for TrueCar, where he is responsible for the company`s overall technology infrastructure and operations strategy. John is a serial entrepreneur and start-up executive with an extensive background building and operating secure and highly scalable technology infrastructure. He was the co-founder and CTO at Preventsys (acquired by McAfee) and led the network penetration testing team for Internet security pioneer Trusted Information Systems. John has been retained as a consultant by numerous world-class technology, financial services and government organizations. Prior to that, John founded one of the first Internet service providers in New York.
Jonathan Gray, CEO & co-founder of Continuuity, is an entrepreneur and software engineer with a background in startups, open source and data. Prior to founding Continuuity, Jonathan was a software engineer at Facebook where he helped drive HBase engineering efforts, including Facebook Messages and several other large-scale projects from inception to production.
Josh Patterson currently runs a consultancy in the big data machine learning space. Previously Josh worked as a Principal Solutions Architect at Cloudera and as a machine learning / distributed systems engineer at the Tennessee Valley Authority where he brought Hadoop into the smart grid with the openPDC project. Josh has a Masters in Computer Science from the University of Tennessee at Chattanooga where he did published research on mesh networks (tinyOS) and social insect optimization algorithms. Josh has over 15 years in software development and is very active in the open source space with projects such as Apache Mahout, Metronome, IterativeReduce, openPDC, and JMotif.
As the Chief Technologist of T-Systems (Deutsche Telekom), Juergen Urbanski is responsible for the development and application of innovative technologies with the goal of growing revenue and profitability. Juergen has a distinctive track record of leading and supporting technology-enabled business transformation at McKinsey, NetApp and Deutsche Telekom. He has also spent 10 years in Silicon Valley serving Microsoft, SAP, Oracle and Symantec on product development issues. He presents this talk in his capacity as a Board Member Big Data & Analytics of the German IT Industry Association BITKOM.
Julian Hyde is an expert in query optimization and in-memory analytics. He is the lead developer of Optiq, the new cost-based optimizer for Apache Hive, an Apache Drill committer, and lead developer of the Mondrian OLAP engine. He is an architect at Hortonworks.
Julien is the lead for Parquet?s java implementation. He also leads Data Processing tool development at Twitter and is on the Apache Pig PMC. His French accent makes his talks attractive.
Justin Sears is a product marketing manager at Hortonworks. He partners with Hortonworks customers to tell stories about how they use Hortonworks Data Platform to create new analytic applications and build modern data architects to solve real-world big data problems. With experience in software and financial services, Justin has led marketing efforts at Wells Fargo, Symantec and multiple startups. Justin’s core professional focus is communicating how marketing technology and data can be applied to create sustained competitive advantage.
Works on realtime streaming platforms like Storm with specific focus on trend detection. Currently working on PIG on Storm and cloud platforms.
Karthik is a member of the platform engineering team at Cloudera, where he works on YARN/MapReduce and other CDH components. He holds an MS and pursuing his PhD at Purdue, focusing on scheduling and resource management for distributed systems.
Karthik Kannan is a successful entrepreneur specializing in product management, marketing and sales. He is currently co-founder and Chief Product Officer of Caspida Inc., an enterprise security startup in the Bay Area. Prior to that, he was co-founder and VP of Products at Cetas, a Big Data analytics company, acquired by VMware in April 2012. The Cetas product line is now a part of Pivotal, a spinoff from EMC and VMware. Prior to Cetas, he was VP of Marketing at Kazeon, a DLP and eDiscovery company, which was acquired by EMC in 2009. He has spoken at various conferences including GigaOm Structure, Data Week 2.0, TiECon etc. More at http://karthikkannan001.blogspot.com/ and @KarthikBigData.
Ken Krugler started using Hadoop back in the dark ages (2006), when it was still part of Nutch. He is the president of Scale Unlimited, a provider of consulting and training services for big data analytics, search, and machine learning using Hadoop, Cascading, Mahout, Cassandra and Solr. Ken is an Apache Tika committer and a member of the Apache Software Foundation, and in his spare time he tries to make Python programming interesting for high school students.
Kevin Minder is the initiator of the Apache Knox Gateway project. Currently he is working at Hortonworks focusing on Hadoop security. He is a veteran in the enterprise middleware space with a deep background in WebServices and manageability. His experience ranges from the creation of the management platform for one of the first Java middleware platforms at Bluestone Software to the development and integration of the WebServices technology used throughout Oracle`s JEE, SOA and application domains.
Kiru Pakkirisamy is an accomplished technology leader with a proven ability to conceptualize and deliver products and solutions targeted at Data-driven enterprises. As a CTO at Serendio, he helps Enterprises harness their everyday data for business advantage through innovative Big data Science solutions. Most recently, he served as a Director of Engineering for Splunk, spearheading the development of Splunk-Hadoop integration technologies. Prior to this, Kiru had held technical and management roles at Successfactors, and Sybase. Kiru?s areas of interest include Distributed Computing, Big Data frameworks, and Predictive Science techniques as applied to Insurance, Healthcare, and Retail.
Kishor Angani Currently working in Video platforms team in Yahoo! We take care of video transcoding and enrichments. Prior to this I have worked on doing abuse analysis on grid for Yahoo! login systems.
Kishore Gopalakrishna is software developer with great passion for using and building large scale distributed systems. As part of Data Infrastructure team at LinkedIn, Kishore has built Espresso, a distributed data store and Helix, a generic cluster management system. Prior to LinkedIn, Kishore spent large part of his time at Yahoo working on Ad systems that mostly involved data analysis using Hadoop and building systems like Apache S4 for near real time stream processing.
KishoreKumar Yellamraju is a Hadoop Operations Engineer at RocketFuel where he builds and manages large scale Hadoop-Yarn Clusters. Kishore works mostly on maintaining , optimizing the Big data Infrastructure components at RocketFuel such as Hadoop, HBase , Kafka, Storm , Hive, Hue, Oozie, Ganglia , OpenTSDB and more. Prior to working at Rocket Fuel, he was a senior systems Administrator at Fiserv. He is a certified Apache Hadoop Administrator, and holds Masters degree in Computer Science and bachelors degree in Electronics and communications.
Koji is an Apache Hadoop&Pig committer involved in supporting hadoop users at Yahoo for over 7 years. He has a Ph.D. in Information and Computer Science from University of California, Irvine.
Kunal Jain is an experienced professional with 12+ years of strong background in Data Warehousing, Database Appliances, Solution Architecture and expert consulting. He has extensive hands-on experience with big data technologies including row based and column based databases like SAP HANA, IBM Netezza, ParAccel and open source databases like Hadoop, Hive, Sqoop and predictive tools like R, SAP Predictive Analysis. Kunal holds a Masters degree in Computer Science from Central Michigan University.
Larry McCay is a veteran in the enterprise middleware space with a focus on manageability and security. He was instrumental in the development of a management platform for an early application server at Bluestone Software. He has participated as an expert on a number of security related JSR`s and helped to shape the security of Java EE. As a member of the Weblogic Security Team at Oracle, he implemented numerous Java EE security features for the application server. He now works as part of Hortonworks` team dedicated to Apache Knox Gateway (incubating) project with a focus on the security aspects of the Gateway and Hadoop core.
@Splunk for 6+ years. Lead for Hunk: Splunk Analytics for Hadoop. Before Hunk, Ledion has been the tech lead for Splunk’s scheduler and alerting, field extractions, saved searches, event types and more. He’s also been responsible for delivering a number of apps that integrate Splunk with external systems: MySQLConnector (superseded by Splunk DB Connect) and Hadoop Connect. You can follow him on twitter @ledbit.
Master of Science in Electrical Engineering from Moscow State Technical University, over 25 years of IT experience, over 15 years of Oracle and RDBMS technologies, 5 years of Hadoop experience. Focusing on system administration, architecting, monitoring tools and customer support.
Lohit graduated from Stony Brook University with Masters in Computer Science. He has been working on Hadoop and related technologies at Yahoo!, Quantcast, MapR Technologies and now Twitter.
Makoto YUI is a researcher of the Information Technology Research Institute at National Institute of Advanced Industrial Science and Technology (AIST), Japan. He is working on a large-scale machine learning as a research project and released Hivemall as an open source software. Find his profile on https://staff.aist.go.jp/m.yui/
Manohar is the Product Manager for several products in large-scale data processing and analysis at Google, including Pregel and MapReduce. These products provide infrastructure support for applications ranging to multi-PB data crunching for Web search. Manohar has an avid, long-running interest in parallel processing. His prior training and experience include a BS from Harvard, two MS degrees and a PhD in Computer Architectures from Stanford. He holds five patents in computer engineering, and his experience spans over two decades of work for Google and HP Labs and as a consultant for start-ups.
Marie-Luce Picard is a project manager and BI expert at EDF Lab. She has managed different R&D projects dealing with business intelligence and information systems (advanced documentation systems, data-mining for customer insight teams, etc..) She has also managed the EDF Lab team working on BI and data analytics. She is currently in charge of managing the EDF Lab project dealing with Big Data to handle the evolutions of EDF information systems linked to the data deluge expected within a few years impacting all businesses of the Company. She is also a member of the Big Data Coordination Committee and of the BI coordination Committee led by the EDF IT Division.
Mark Tomallo, MBA/CISSP, is the Director of Information Security and Risk for Cardinal Health in Columbus, Ohio. Mark has responsibility for the protection of Cardinal Health’s digital assets and the calculation/communication of IT risk throughout the organization. Mark’s background includes positions within Cisco Systems as well as CEO roles at The James Group and The Vigilance Group, security consulting companies he founded with the former being sold to AT&T EasyLink. He has worked closely with executive level decision makers in defining their technology and organizational goals pertaining to compliance, information security, risk, and emergency response for more than 18 years.
Principal Scientist with Microsoft`s Cloud and Information Services Lab. There, I work on large scale machine learning and datamining applications, algorithms and most prominently the systems to run them on.
Martin is an engineer in the Data Infrastructure team at Facebook working on a query engine for interactive data analysis. Previously, he was at Proofpoint and Ning, where he designed and built large distributed systems. His past experience spans many areas, including computer languages, compilers, databases, search, machine learning and graphics.
Matt Brandwein is Director of Product Marketing at Cloudera. Prior to Cloudera, he delivered enterprise search, business intelligence, and data discovery products at Endeca and Oracle. Matt holds degrees in Computer Science and Mathematics from the University of Massachusetts Amherst.
Matt is a Principal Program Manager on the Azure HDInsight (Hadoop) team at Microsoft focused on building tools to make developers’ lives easier. Matt has been at Microsoft for 8 years working on developer tooling for distributed systems and cloud applications as well as technical evangelism.
Matthew Farrellee is a Software Engineer in the CTO office at Red Hat with over a decade of experience in distributed and computational system development and management. Matt has been involved with numerous open source projects over the years. His current focus is on big data technologies, including combining OpenStack and Hadoop through the Savanna project. He is also active in the Fedora Big Data SIG and the Fedora community in general.
Mayank Bansal is an Apache Hadoop Committer and Apache Oozie PMC and committer. He has been working on Hadoop and Oozie more than 4 years previously from Yahoo! and now from Ebay, Inc.
Michal is the Tech Lead for the Pregel project, on which he?s worked since joining Google in 2011. Michal received his PhD from the University of California, Santa Barbara, where he conducted research on programming language design and implementation, improving memory management performance, interactions between virtual machines for different languages, and type-safe object sharing for co-located VMs for statically- and dynamically-typed languages. He?s been on the external review committee for the ASPLOS, PLDI, PPoPP and TPDS conferences. He?s also previously worked for Sun Microsystems and Motorola.
Mithun Radhakrishnan is a committer on the HCatalog project, and a Hive developer at Yahoo. He’s the author of DistCp on Hadoop 0.23+. He’s an erstwhile firmware developer and is prone to flare-ups from C++ withdrawal.
Senior Principal in Y! working on Big Data platforms specifically streaming and related applications for realtime trend detection. Also works on content extraction & enrichments using Machine Learning.
Naresh has extensive experience and deep interest in large scale data management domain & infrastructure. Currently, Naresh is managing the data platform engineering group at Inmobi
Nick found Hadoop when his nightly ETL jobs started taking 20+ hours to complete. Since then, he has applied Hadoop and HBase to projects over social media, social gaming, click-stream analysis, climatology, and geographic data. Nick also helped establish Seattle?s Scalability Meetup and tried his hand at entrepreneurship. He is an HBase committer and coauthored “HBase in Action,” the unofficial user`s guide for HBase. His passion is scalable, online access to scientific data.
After a Master degree in distributed real-time systems, Nicolas has stayed focused on the software architecture business at various positions including Head of Architecture at Thomson Reuters for the Risk Management product line. He has been deeply part of the Big Data arena for more than 2 years, working especially with Hortonworks on HBase MTTR. He combines traditional software and enterprise architecture skills with a deep knowledge of Big Data architecture. Nicolas is PMC member for the Apache HBase project. He is also cofounder of Scaled Risk, a company that provides a Big Data solution on top of Hadoop and HBase.
Nong Li is a software engineer working on the open-source Cloudera Impala. He spends most of his time focusing on improving the performance of the query execution engine, working on the IO subsystem, JIT-compiling portions of the query execution, and working on expression evaluation and other performance-centric components.
Ofer Mendelevitch is Director of data sciences at Hortonworks, where he is responsible for professional services involving data science with Hadoop. Prior to joining Hortonworks, Ofer served as Entrepreneur in Residence at XSeed Capital where he developed an investment strategy around big data. Before XSeed, Ofer served as VP of Engineering at Nor1, and before that he was Director of engineering at Yahoo! where he led multiple engineering and data science teams responsible for R&D of large scale computational advertising projects including CTR prediction (with Hadoop), a new front-end ad-serving system and sales tools.
Oleg Checherin is an independent consultant in area of Hadoop implementation and integration with corporate systems. He has Sc.M. in Electrical Engineering from Moscow Power Engineering Institute and over 20 years of IT experience with primary focus with data support in enterprise scale.
Oscar Boykin (@posco) is a member of the analytics infrastructure team at Twitter and committer on scalding, algebird, summingbird and several other Twitter open source libraries.
Oscar Padilla leads the strategic services for Luminar’s analytic and technology solutions. He is accountable for formulating new product and service offerings, as well as helping clients achieve higher efficiency from both analytic and technology solutions across online and offline media. With nearly 20 years of digital media under his belt, Oscar has worked with Hitachi Consulting, Blackboard Inc., RepEquity (previously Mindshare Interactive) and Bridgeline Digital. Oscar has provided expert perspective for industry trade publications and has been published in Mobile Marketer, DMA News, 1to1 Magazine, Value Retail News, Chief Marketer and iMedia Connect.
Owen O`Malley is a cofounder and architect at Hortonworks, a rapidly growing company that supports customers using Hadoop. Owen has been working on Hadoop since the beginning of 2006 and was the first committer added to the project. In the last 8 years, he has at various times been the architect of MapReduce, Security, and now Hive. Before working on Hadoop, he worked on Yahoo Search`s WebMap project, which was the original motivation for Yahoo to work on Hadoop. Prior to Yahoo, he wandered between testing (UCI), static analysis (Reasoning), configuration management (Sun), and software model checking (NASA). He received his PhD in Software Engineering from University of California, Irvine.
Taylor is an Apache Storm committer and release manager with over 17 years of software development experience in various sectors including healthcare, DoD, and transportation logistics. Prior to joining Hortonworks, he worked at Health Market Science where he led the integration of Storm into HMS’ next generation Master Data Management platform with technologies such as Cassandra, Kafka, Elastic Search, and the Titan graph database.
Professor Doctor Peter Boncz is a senior research scientist at CWI and part-time lecturer (UHD) at VU University Amsterdam, specializing in database system architectures. He architected two database systems, MonetDB and X100 (aka VectorWise). He also participated in two successful CWI spin-off companies in large-scale data management: Data Distilleries and VectorWise. He is internationally well-known for work on columnar database systems, XML data processing and architecture-conscious database research. MonetDB and VectorWise are widely regarded as state-of-the-art analytical database systems.
Peter Guerra is a Principal in Booz Allen Hamilton’s Strategic Innovation Group leading a large team of Data Scientists. He has 15 years of professional experience applying computer science servicing National Intelligence, Military, Commercial Health, and Financial Services clients. His specialty is in highly available, large-scale distributed systems and advanced analytics, and is responsible for leading several large-scale Hadoop computing projects. He has been a software and security consultant to government and commercial organizations throughout his diverse IT career, focusing on software development, security engineering, and highly available system design.
Prafulla Wani is a big data engineer with experience in Hadoop/NoSQL based solutions as well traditional data-warehouse implementations using relational databases and ETL tools. With more than 10 years of IT experience and majority of it has been working at client locations in US, he is currently playing role of Technical Architect – Big Data as part of Strategic Offering Group (SOG) at Syntel.
Rahul Ravindran is a software generalist with extensive experience in backend systems. Currently, he is a data engineer at BrightRoll, the largest independent video ad technology platform, where he uses Hadoop, Hbase, Flume, MapReduce to solve some of the advertising industry’s most complex challenges. Prior to Brightroll, Rahul worked at Zynga, helping scale and improve performance of the Farmville backend during its hypergrowth, and previously at Microsoft where he wrote device software for Xbox, Courier tablet, Zune devices. Rahul has a BE, Computer Science and Engineering from R. V. College of Engineering in Bangalore, and an MS in Computer Engineering from the University of California, Santa Cruz.
Raj Nair is currently the Director of Data Management and Engineering at Penton Media. At Penton, Raj is focused on building a scalable data management platform combining both SQL and NoSQL technologies, a platform that would ultimately help create new content-centric products. Prior to Penton, Raj was at EMC innovating in the areas of Risk, Compliance, and Governance products, Earlier at IBM/Informix, Raj worked on Database drivers and performance accelerators. With over 15 years of technical and management expertise in High Performance Databases, Nosql and Hadoop technologies, Raj has the unique ability to combine customer needs and emerging technologies into innovative products.
Remy is a Software Architect and IT Consultant at Octo Technology. After his studies at Epitech Paris where he first worked on distributed systems for raytracing rendering and designed and implemented an exokernel for the School?s System Lab (LSE), he has been involved in several IT projects, working with various technologies such as Opensource ECM and ERP, electronic Strongbox, iPad based video recommandation, distributed systems for simulation or computing and Hadoop for ETL, BI and Machine Learning workload. He also taught software development patterns at ETNA, Paris. He is currently in charge of Octo?s Hadoop R&D and co-organizer of the Paris Hadoop User Group.
Chief Information Officer Reuben Shaffer oversees QuestPoint’s architecture, infrastructure and data – from data collection to aggregation and manipulation. In his role, he ensures the 11 billion impressions and 5 billion requests per day that QuestPoint receives on average are seamlessly executed through the company’s data intelligence platform. Through Shaffer’s leadership, companies are able to mine consumer behavior data, affording them powerful competitive advantages for their business’ bottom lines.
I am a Software Engineer at Facebook. I work on building data storage systems on HBase for various applications at Facebook. Before joining Facebook, I worked at NetApp for almost 2 years in the data mobility team. I have done my Bachelors in Information Technology from VJTI, India and Masters in Computer Networking from Carnegie Mellon University, USA.
Rohini currently leads Pig and Oozie development at Yahoo!, and has been working on Hadoop and related projects like Pig, Oozie, HCatalog, Hive, Grid Data Lifecycle Management for the past 5 years at Yahoo! scale. Rohini is a PMC member/committer on the Apache Pig project, and a committer on the Apache Oozie project. She is interested in large-scale data processing and is currently working on Pig-on-Tez which targets low latency ETL on Hadoop.
Roopesh Varier leads the Big Data Platform team in Symantec Cloud Platform Engineering Group. Previously, he led the development of a big data platform in Symantec’s Threat analysis organization – building it from scratch to become the largest known security metadata store in the world. Prior to that he has also held multiple leadership roles in other companies.
Russ is the VP of Data Platform at TrueCar.com, where he creates the intelligence systems driving TrueCar?s innovative interactive product set. Prior to TrueCar, he held executive, product and technical leadership positions at category leaders like IAC, Grind Networks, and Wolfram|Alpha. Russ holds a degree in mathematics from the University of Chicago and currently lives in Marina Del Rey, CA with his wife and two daughters.
Sagi has over 20+ years experience in software engineering from Web commerce & front-end development to large-scale data processing and virtually everything else in between. He currently oversees Yahoo’s grid services and infrastructure, which includes the entire spectrum of Hadoop implementations. As part of Yahoo’s Hadoop team, his role includes hands-on engineering and development of tools to improve Hadoop’s services, and to help shape the future of the infrastructure. His team maintains all of Yahoo¹s grid machines, which include some of the largest YARN clusters in the world.
Sameer Agarwal is a Ph.D. candidate in the AMPLab at Berkeley working on large-scale approximate query processing frameworks. His research interests are at the intersection of distributed systems, databases and machine learning. He received his B.Tech in Computer Science and Engineering from the Indian Institute of Technology, Guwahati and was awarded the President of India Gold Medal in 2009. He was supported by the Qualcomm Innovation Fellowship during 2012-13 and is supported by the Facebook Graduate Fellowship during 2013-14.
Samuel currently works at Cisco Systems in the Application Centric Infrastructure group. Samuel’s prime focus areas are application profiles from a network perspective and network programmability. He is one of the co-authors of “Big Data in the Enterprise“. His interests include low latency networks, SDN, monitoring, management and automation.
Sanjay is founder and architect at Hortonworks, and an Apache Hadoop committer and member of the Apache Hadoop PMC. Prior to co-founding Hortonworks, Sanjay was the chief architect of core-Hadoop at Yahoo and part of the team that created Hadoop. In Hadoop he has focused mostly on HDFS, MapReduce schedulers, high availability, compatibility, etc. He has also held senior engineering positions at Sun Microsystems and INRIA, where he developed software for distributed systems and grid/utility computing infrastructures. Sanjay has a PhD in Computer Science from the University of Waterloo in Canada.
Santosh is a software executive with 25 plus years of software development experience. Prior to founding Aziksa, Santosh was CTO for Kovim, a Bay Area learning solution company serving global enterprises. Before Kovim, Santosh was VP of Product Development at GlobalEnglish, where he successfully led the team to build multiple learning products with SaaS model with cloud hosting. Expert in the building strategy, planning, design and delivery of cost effective, high performance technology solution in support of company growth.
Saran has over 10 years of experience in digital analytics and eCommerce strategy, implementation and management. He specialises in extracting various forms of online and social media data and integrating it with offline data, and utilising for single customer view reporting and predictive modelling.
Saravanan holds Master of Computer Applications from Anna University and PG. Diploma in Cyber Crime Investigation and Forensic from Asian School of Cyber laws. Hard core developer with 5 years of experience, having versatile programming knowledge. Involved in Research & Development of Big data stack
Savin is a Rocket Scientist at Rocket Fuel in the modeling infrastructure team working on large scale data pipelines for data mining and machine learning. He graduated from Indian Institute of Technology, Delhi with a BS in Computer Science.
Scott Gray is a senior architect for IBM’s InfoSphere BigInsights Big SQL solution. Scott has an extensive career in the computer industry focusing heavily on relational database, architecture, design, optimization and internals. Prior to working with IBM, Scott was the chief architect for ANTs Software’s SQL Skin for Sybase, a real time Sybase T-SQL to IBM SQL PL translation engine.
Out of the CTO Office, Scott works across all of Red Hat’s business units to shape Red Hat’s overall “Big Data” strategy. At the same time, Scott manages the “Big Data” engineering team which has focus (so far) on three areas: Hadoop connectivity and compatibility for Red Hat’s storage solution(s), elastic provisioning and workload management of Hadoop for OpenStack, and incorporation of upstream BigData projects into Fedora. Red Hat’s BigData initiatives continue to broaden, with much of that development coming from the BigData Engineering team out of the Red Hat CTO Office.
Sheetal is a Principal Architect working with Hortonworks. He has strong expertise in Hadoop ecosystem with very rich & diverse field experience across various verticals including Telco, Hi Tech, Retail, Internet Companies etc. He has served in key positions as Lead Big Data Architect, SOA Architect in variety of extremely large & complex enterprise programs. Has extensive knowledge of BigData/NoSql technologies including Hadoop/Yarn/Hive/Pig/HBase/Storm/Kafka/ElasticSearch etc. He has defined & established data architectures for multi-petabyte warehouses on Hadoop, has extensive hands on experience in deploying, tuning very large Hadoop clusters & building scalable applications on them.
Shital Mehta Currently an Architect in Video team where I work on video transcoding and enrichments platform. I have also worked in advertising domain where my main focus was on abuse detection. Before Yahoo! I spent considerable time in VoIP domain building voice mail, video mail, audio/video/web conferencing solutions for Telecom and Cable service providers.
Shivaram Venkataraman is a third year PhD student at the University of California, Berkeley and works with Mike Franklin and Ion Stoica at the AMP Lab. He is a committer on the Apache Spark project and his research interests are in designing frameworks for large scale machine-learning algorithms. Before coming to Berkeley, he completed his M.S at the University of Illinois, Urbana-Champaign and worked as a Software Engineer at Google.
Siddharth Wagle is a Member of Technical Staff at Hortonworks and a committer and PMC member for the Apache Ambari project. His primary focus is developing Ambari backend for provisioning, managing, and monitoring Apache Hadoop clusters. His previous background is in building high performant and scalable systems and APIs at Telenav Inc and Intelliun Corp.
Chandra leads the infrastructure areas used for machine learning systems at Rocket Fuel. Prior to Rocket Fuel, he worked on peta byte scale Stats systems in the Ads Infrastructure team at Google and querying/indexing of semi structured data in the database kernel at Oracle. Chandra has a M.S in Computer Science from the University of Wisconsin, Madison and holds 100+ patents in database systems.
Hadoop developer at Syntel with over 3+ years of experience in IT industry with extensive experience in Hadoop and NoSQL based solutions.
Srikanth Sundarrajan works at Inmobi Technology Services, helping architect and build their next generation data management solution. He has been actively involved in various projects under the Apache Hadoop umbrella including HDFS, MR, Oozie. He has been working with distributed processing systems for over a decade and Hadoop in particular over the last four years. He was with the Hadoop team earlier at Yahoo!.
Srimanth Gunturi is an Apache Ambari committer and PMC member working at Hortonworks.
Srinivas Nimmagadda is responsible for the architecture and technology of petabyte scale big data analytics cloud platform at Symantec’s Cloud Platform Engineering team. Prior to Symantec, Srinivas has led the efforts in building a large-scale private cloud platform at Intuit while pioneering the use of Software Defined Data Center (SDDC) concepts. Earlier in his career, Srinivas was instrumental in developing one of the world`s largest grid computing platforms for HPC workloads running on both Unix and Windows environments.
Stefan Groschupf is a big data veteran and serial entrepreneur with strong roots in the open source community. He was one of the very few early contributors to Nutch, the open source project that spun off Hadoop, which 10 years later, is considered a 20 billion dollar business. Stefan is currently the CEO of Datameer, the first big data analytics tool built natively on Hadoop.
As Rante`s chief operating officer, Stephanie Caprini leads the company?s global sales, marketing and services organization. Under her leadership, she oversees worldwide sales, field marketing, services, support and partner channels, and corporate support functions including Information Technology, Worldwide Licensing & Pricing and Operations. The sales and marketing organization is focused on delivering Rante`s software and services to customers and partners all over the world. At Rante, Stephanie has driven a strong track record of results, execution excellence and improved efficiency while also driving the customer satisfaction scores to the highest in company history.
Steve has led sales organizations at numerous technology companies, most recently at Packet Design, where he was EVP, WW Field Ops. Steve led Packet Design’s field for over 6 years until acquisition. Prior to that, he was VP WW Sales at MonoSphere (Dell), Satmetrix, and Vicinity (Microsoft). Steve was an early member of Marimba (BMC) where he built their reseller and direct channels as regional VP Sales. Before that, he held sales management positions at Tivoli (IBM) and Sun (Oracle). Steve received his J.D from the Santa Clara University School of Law and his B.S. from the San Diego State University School of Business.
Staff Engineer at Inmobi working on an in-house Data warehouse platform that facilitates querying and managing large datasets residing in Hadoop. Earlier to this worked In Yahoo on Distributed Data processing and Ingestion platforms which leveraged Hadoop.
Sumeet Singh is a Senior Director of Products at Yahoo responsible for platforms product management and customer engagements. In this role, he also leads the Hadoop products team responsible for both Apache open source contributions and Yahoo projects. Sumeet has 15 years of Product Management, Product Development, and Strategy Consulting experience in the technology industry. Sumeet earned his MBA from UCLA Anderson School of Management and MS from Rensselaer Polytechnic Institute, NY.
Sumit Mohanty works at Hortonworks on the Apache Ambari project. He is an Apache Ambari Committer and PPMC member. He is currently working on managing long lived YARN application. Prior to joining Hortonworks he has worked for several years at Microsoft on various aspects of system management and monitoring. He holds a PhD from University of Southern California.
Sunil works in the ads and data team in Yahoo! Areas of current focus is to build large scale systems for distributed analytics.
Supreeth works in the ads and data team in Yahoo! Areas of current focus is to build large scale systems for distributed analytics.
Suresh is an Apache Hadoop committer and member of Apache Hadoop Project Management Committee (PMC). He is a long term active contributor to the Apache Hadoop project and has designed and developed many significant features for Hadoop. Prior to co-founding Hortonworks, he served as a software architect at Yahoo! working on Apache Hadoop, where he developed features and supported some of the largest installations of Hadoop clusters. Follow Suresh on Twitter: @suresh_m_s
Tanya is co-founder of Natero, a big data cloud analytics company. Prior to Natero, as the Big Data lead at Google she led big data partnerships and ecosystem strategy for the Google Cloud Platform. In the Office of the CTO at NetApp she initiated and led the development of the company`s big data product and partner strategy. She started her career at Cisco where she held various roles in product management and software development. Tanya has an MBA from Haas, UC Berkeley and an MS in Electrical Engineering from SUNY Stony Brook, NY.
Blending both industrial and academic research, Tatsiana is expert at solving hard business problems. She brings a background in both mathematics and statistics, and has deep experience researching and implementing models for predicting user behavior.
Ted is Chief Application Architect for MapR Technologies and contributes to several Apache open source projects including Mahout Hadoop, Zookeeper and Hbase. He is also a mentor for Apache Drill and Apache Storm. Ted has a Ph.D. in computing science from the University of Sheffield and is named as inventor on 24 issued patents with a dozen more pending. He also bought the beer at the first Hadoop User Group meeting.
Thiruvel Thirumoolan is a developer in the Hive and HCatalog team at Yahoo!. In this role he is responsible for deployment of Hive, HiveServer2 and HCatalog across all the Hadoop clusters at Yahoo! and ensuring they work at the scale for the usage patterns of Yahoos. He also contributes the features and fixes to the Apache Hive community. He has a Bachelors degree from Anna University and has been working in Hadoop team at Yahoo! for more than 4 years. His favorite theme at Yahoo! internal Hack Days is Hadoop and also mines the trove of Hadoop logs for usage patterns and insights.
Thomas Graves is a Principal Engineer at Yahoo. He is a Hadoop committer and PMC member, a Spark committer and Incubator PMC member at the Apache Software Foundation.
Timothy Los is the Chairman and CEO of Rante. He is also the Chairman of Interfint. Rante is engaged in cloud procurement and database management and synchronization services. Interfint is involved in FATCA-related solutions for governments and financial institutions. Mr. Los attended St. John`s School of Law. Mr. Los is currently engaged in NYLS Taxation LLM with a focus on international taxation, transfer pricing, and cross-border legal and taxation issues with the exchange and use of intellectual property. At Rante, Mr. Los oversees the daily business operations and is in charge of the growth of Envizn is current and new markets.
As VP of Product Management, Tomer is responsible for product strategy and roadmap. Prior to MapR, Tomer held product management and engineering roles at Microsoft, most recently as the product manager for Microsoft Internet Security & Acceleration Server (now Microsoft Forefront). He is the founder of two websites that have served tens of millions of users, and received coverage in prestigious publications such as The New York Times, USA Today and The Times of London. Tomer also authored a 900-page programming book. He holds an MS in Computer Engineering from Carnegie Mellon University and a BS in Computer Science from Technion – Israel Institute of Technology.
Tony Baer leads Ovum’s Big Data research area. Over his 25 years in the industry, he has studied issues of data integration, software and data architecture, middleware, and application development. Having tracked the emergence of BI and data warehousing back in the 1990s, Baer sees similar parallels emerging in the world of Big Data today. His coverage focuses on how Big Data must become a first-class citizen in the data center, IT organization, and the business.
For the past 18 years, Tony has been focused on helping organizations maximize the impact of their Information Integration and Governance objectives. Prior to joining the IBM product management team in 2009, Tony spent many years building scalable data integration solutions using InfoSphere DataStage’s massively parallel data integration capabilities. Tony now focuses attention on the IBM roadmap for investments in the InfoSphere Information Server platform, and applies specific attention to how organizations can fully exploit the unique capabilities of Hadoop as part of their modern data management platform.
Before Hulu, Tristan was VP of Solution Design at Ares Mgmt, leading a team building research tools for investment professionals. He has taught software development courses for IBM, BEA Systems and others in the US, Europe and Asia. He received his CS degree from Duke University, and completed coursework in Applied Mathematics and various data analysis at Columbia U., Suffolk U. and U. of Illinois at Urbana-Champagne. He earned the CFA designation in 2006. Previously, Tristan built risk management and data analysis tools at Capital Research as a Quant Research Associate. Before his career in finance, he participated in a number of start-ups, both as a resource and as principal.
My research aims to provide appropriate data management toolkits for the data scientist (machine learning practitioners) to operate at scale. I am currently involved with developing a declarative language for Big Data Analytics that targets machine learning and graph-based algorithms.
Mr Varun Vasudev was Senior Sofware Engineer at Yahoo!, and is experienced in web search. Now he is a Member of Technical Staff at Hortonworks, Inc. He is actively contributing to Hadoop YARN project.
Venkatesan Ramachandran is a member of the Analytics Infrastructure team at LinkedIn helping to architect and build next generation distributed data processing and analytics platforms and solutions. He has been working with distributed processing systems and analytics for over a decade with a focus towards Hadoop related technologies in last 5 years. Previously he was part of the Yahoo Hadoop team architecting and building Grid Data Management Platform and business critical data processing pipelines and infrastructure.
Seetharam Venkatesh works at Hortonworks Inc. leading the data integration development efforts. He is an active contributor to Apache Oozie, Apache Sqoop and Apache Flume. He was part of the Hadoop team at Yahoo where he built data management solutions. He has been involved in Hadoop and contributed to many open source projects in the ecosystem over the last 6 years.
Vikram Dixit is an active committer on the Apache Hive project and has been contributing to other Apache projects over the past 2 years at Hortonworks. He is also a committer on the Apache Tez and Ambari projects. At Hortonworks, he is involved in the design and development of Hive as part of the Stinger initiative delivering performance and SQL capabilities in the ecosystem. Vikram holds a Masters degree in Computer Science from the University of Southern California.
Vinod Kumar Vavilapalli is the Hadoop YARN and MapReduce guy at Hortonworks. He is a long term Hadoop contributor at Apache, Hadoop committer and a member of the Apache Hadoop PMC. He has a Bachelors degree from Indian Institute of Technology Roorkee in Computer Science in Engineering. He has been working on Hadoop for more than 6 years and he still has fun doing it. Straight out of college, he joined the Hadoop team at Yahoo! Bangalore where he worked on HadoopOnDemand, Hadoop-0.20, CapacityScheduler, and Hadoop security, before Hortonworks happened. He is passionate about using computers to change the world for better, bit by bit. He is reachable at twitter handle @tshooter.
ScaleOut Software was founded in 2003 by Dr. William L. Bain, who earned a Ph.D. (1978) in electrical engineering/parallel computing from Rice University and has worked at Bell Labs research, Intel, and Microsoft. Dr. Bain founded and ran three start-up companies prior to joining Microsoft. In the most recent company (Valence Research), he developed a distributed Web load-balancing software solution that was acquired by Microsoft and is now called Network Load Balancing within the Windows Server operating system.
Xiangrui Meng is a software engineer at Databricks, the company founded by the creators of Spark. His main interests center around developing and implementing scalable algorithms for scientific computing. He has been actively involved in the developments of Spark MLlib since he joined, contributing new features and helping review pull requests. Before Databricks, he worked as an applied research engineer at LinkedIn, where he was the main developer of an offline machine learning framework and provided user support to multiple offline recommendation pipelines.
Xuan works as a Software Engineer at Hortonworks, where he focuses on the Hadoop Yarn and MapReduce. Prior to this, he spent a year and half working as a Software Engineer on big data processing at ADP.
Yanpei is a member of the Performance Engineering Team at Cloudera, where he works on internal and competitive performance measurement and optimization. His work touches upon multiple interconnected computation frameworks, including Hadoop, Impala, HBase, Search, and Hive. He holds a Ph.D. from UC Berkeley.
Dr. Zhijie Shen was awarded a Ph.D. degree in Computer Science from National University of Singapore. Now he is a Member of Technical Staff at Hortonworks, Inc. He is a Apache Hadoop Committer, and one of the core team of Apache Hadoop YARN. Moreover, he has been actively contributing to Hadoop ecosystem since 2011.