Prior to co-founding Hortonworks, Eric served as VP Hadoop Software Engineering for Yahoo!, where he led the evolution of Apache Hadoop from a 20 node prototype to a 42,000 node service that is behind every click at Yahoo!. Eric also served as a technology leader for Inktomi’s web service engine, which Yahoo! acquired in 2003. Prior to Inktomi, Eric developed software for video games, video post production systems and 3D modeling systems. Eric has a Master’s degree in Computer Science from the University of California, Berkeley and a Bachelor’s degree in Mathematics and Computer Science from Carnegie Mellon University. Follow Eric on Twitter: @jeric14.
As Research Director for data management and analytics within 451 Research’s Information Management practice, Matt has overall responsibility for the coverage of operational and analytic databases, data integration, data quality, and business intelligence. Matt’s own primary area of focus is on relational and non-relational databases, data warehousing, data caching, and Hadoop. Matthew is also an expert in open source software and regularly contributes to 451 Research’s open source-related research. Prior to joining 451 Research, Matthew was Deputy Editor of monthly magazine Computer Business Review and ComputerWire’s daily news service. Matthew started in IT journalism in 1997 as a staff writer with monthly magazine Unix and NT News, where he became editor in 2000. He has also contributed to a number of other publications. Matthew holds a BA in Multimedia Journalism from Bournemouth University in the UK.
Shaun has more than 20 years of experience in the software industry, with a track record of building early stage and midsize software companies into successful market leaders. Shaun was previously VP of Product Strategy at VMware where he focused on defining VMware’s cloud application platform vision. Shaun has also held VP and Director level positions at SpringSource, Red Hat, JBoss, Princeton Softech, HP, Bluestone Software, and Primavera Systems. He holds a B.S. in Electrical Engineering from Drexel University. Follow Shaun on Twitter: @shaunconnolly.
Abhijit Lele works as Solutions Engineer at Hortonworks.
As a member of the CTO Office, Alan is responsible for design and development of innovative solutions using emerging technologies in data analysis and machine learning. Alan is also currently studying Computer Engineering at the University of Ottawa.
Alan is an original member of the engineering team that took Pig from a Yahoo! Labs research project to a successful Apache open source project. Alan also designed HCatalog and guided its adoption as an Apache Incubator project. He is a co-founder of Hortonworks, where he now works on Pig, Hive, and HCatalog. Alan has a BS in Mathematics from Oregon State University and a MA in Theology from Fuller Theological Seminary. He is also the author of Programming Pig, a book from O?Reilly Press.
Alasdair Anderson, Head of HSS IT Architecture, HSBC. Alasdair is currently responsible for the overall technical architecture at HSBC Securities Services having previously been the Lead Data Architect for the division. As part of the HSBC Global Banking and Markets Innovation programme Alasdair is also responsible for leading the investment banks strategy on Big Data platforms. Prior to joining HSBC in 2008 Alasdair worked as an independent consultant for 10 years, clients included JP Morgan, BNP Paribas, Fitch Ratings, Man Investments and the Bank of Bermuda.
Alex Gorbachev is the CTO of Pythian leading the office of innovation. Alex has been working in the data industry for 15 years and architected multitude of business critical data-driven solutions. In the last few years, he`s been a frequent presented at dozens of data conference around the world.
Allen Wittenauer is currently the Senior Grid Computing Architect for LinkedIn, Inc. He has been working with Hadoop with an eye towards operability for nearly six years.
Amit Rustagi is Principal Architect at eBay, and leads the design and strategy of its analytics and experimentation infrastructure. Before eBay, Amit was Sr. Principal architect at Yahoo!. Prior to Yahoo!, Amit held lead role while working on Oracle Applications at Oracle Corp. Amit received a BS in Electronics and Communications from India
Arun is the lead of the MapReduce project in Apache Hadoop where he has been a full-time contributor to Apache Hadoop since its inception in 2006. He is a long-time committer and member of the Apache Hadoop PMC and jointly holds the current world sorting record using Apache Hadoop. Prior to co-founding Hortonworks, Arun was responsible for all MapReduce code and configuration deployed across the 42,000+ servers at Yahoo!. In essence, he was responsible for running Apache Hadoop?s MapReduce as a service for Yahoo!. Twitter: @acmurthy. He is directly responsible for every bit of code and configuration of Map-Reduce deployed at over 40,000 machines running Apache Hadoop.
Arun is the lead of the MapReduce project in Apache Hadoop where he has been a full-time contributor to Apache Hadoop since its inception in 2006. He is a long-time committer and Chair of the Apache Hadoop PMC and jointly holds the current world sorting record using Apache Hadoop. Prior to co-founding Hortonworks, Arun was responsible for all MapReduce code and configuration deployed across the 42,000+ servers at Yahoo!. In essence, he was responsible for running Apache Hadoop?s MapReduce as a service for Yahoo!. Twitter: @acmurthy. He is directly responsible for every bit of code and configuration of Map-Reduce deployed at over 40,000 machines running Apache Hadoop.
Bikas has been working on Apache Hadoop for over a year and is a committer on the project. He has been a key contributor in making Hadoop run natively on Windows and has focused on YARN and the Hadoop compute stack. Prior to Hadoop, he has worked extensively on the Dryad distributed data processing framework that runs on some of the worlds largest clusters as part of Microsoft Bing infrastructure. @bikassaha
Chris Hillman is Principal Data Scientist in the International Advanced Analytics team at Teradata. He has 19 years’ experience working in the Business Intelligence industry, mainly in the Retail and CPGN vertical, working as Solution Architect, Principal Consultant and Technology Director. Chris works with the Teradata-Aster Centre of Excellence and is involved in the pre-sale and start-up activities of Teradata Aster projects, helping customers to understand whether MapReduce or SQL is an appropriate technique to use. Chris is currently studying part-time for a PhD in Data Science at the University of Dundee applying Big Data analytics to the data produced from experimentation into the Human Proteome.
Clemens Neudecker, M.A., Technical Coordinator for European projects in the Innovation and Development department of the KB National Library of the Netherlands. From 2003 – 2009 he also worked at the Bavarian State Library where he was involved in numerous digitization projects. Recently working on large scale projects IMPACT (www.impact-project.eu) and SCAPE (www.scape-project.eu), he has developed a particular interest in OCR and scalable digitization workflows.
Dan Rosanova Microsoft Integration Most Valuable Professional with over fourteen years of experience delivering solutions on Microsoft and Solaris platforms in the financial services, insurance, banking, telecommunications, and logistics industries, where he has specialized in high volume and low latency distributed applications. Dan?s recent focus has been on evolutionary computation, Hadoop, and GPU computing. Dan is the author of Microsoft BizTalk Server 2010 Patterns Packt Publishing. Dan is a senior architect in the Technology Integration practice at West Monroe Partners, an international, full-service business and technology consulting firm.
Graduated from the University of A Coruna (Spain) in Computers Engineering with a Masters Thesis about efficient communications in Java. I joined Yahoo! as a research engineer working on the Scalable Computing group in Barcelona, where I work in distributed transactions processing and hardening of distributed systems against data corruptions.
Eric is leading efforts to employ Hadoop technologies and provide utilities and their customers low-latency access to Opower’s energy efficiency data and analysis that has to date saved more than 2 terrawatt-hours of energy.
Dr. Eric Hanson is a Program Manager Architect in the Big Data team in Microsoft SQL Server, where he focusses on improvements to Hive query performance. He initiated the work that lead to the columnstore index feature that shipped in SQL Server 2012. Prior to coming to Microsoft in 2002, he was a computer science professor at the University of Florida for 10 years, and founded a database alerting software company. He received his bachelor?s degree in computer science from Cornell in 1983 and his PhD in computer science from UC Berkeley, specializing in database management systems, in 1987.
Erich Hochmuth is the Data & Analytics Architecture Lead in Monsanto R&D IT. Since 2009 he and a cross functional team have been working to allow Monsanto to become an early adopter of Big Data technologies. Erich has lead projects that apply technologies such as Hadoop, HBase, and Lucene to data sets from genomics to geospatial.
FLAVIO JUNQUEIRA is a Senior Research Scientist with Yahoo Research and leads the Scalable Computing group in Barcelona, Spain. He holds a PhD degree from University of California San Diego (UCSD) in computer science. He is an active contributor of Apache project such as ZooKeeper, BookKeeper, S4, and HBase. He is also a committer of ZooKeeper, BookKeeper, and S4, and a PMC member of ZooKeeper.
George Trujillo is an internationally recognized speaker and leader in the Big Data, Cloud, Virtualization and Oracle ecosystems. George has been a VMware Tier One Oracle/BCA specialist for the Fortune 200 and strategic accounts. Additional areas of expertise include Hadoop, Java, Linux, Oracle Database and Oracle Fusion Middleware. George has been a Data Specialist for years helping customers implement best practices, proven design patterns and reference architectures for enterprise deployments. Industry recognitions include:Oracle Double ACE, Sun Ambassador for Sun Microsystem`s Application Middleware Platform, VMware Certified Instructor, MySQL`s Socrates Award and MySQL Certified DBA.
Gopal Vijayaraghavan has spent many years making PHP faster at Yahoo & Zynga and thinks about performance problems as opportunities to contribute. He has turned towards Hadoop for newer challenges in a very different cloud & is chasing performance at both ends of the scale spectrum. Beyond work, he’s a photographer by day & biker by sunset and by night posts random thoughts as @t3rmin4t0r
Gunther Hagleitner is a key contributor to Hortonworks’ Hive team. He has been working on and with database technologies for more than a decade and has been active in the Hadoop ecosystem for more than 3 years.
Ivan de Prado is the CEO of Datasalt, a company based in Barcelona which offers Hadoop-based consulting and which is building innovative open-source products around the Hadoop eco-system such as Pangool or Splout SQL. Iv_n has worked on several start-ups such as Strands and Enormo in a wide variety of roles, posing several challenges and therefore gaining broad experience in the Big Data space. He has co-authored a paper on extending MapReduce which has been accepted and will be presented in ICDM2012.
Ivan Provalov is a software developer and architect with over 15 years of professional experience in software architecture, design and development. He currently holds a research engineer position at LucidWorks developing LucidWorks Big Data product. Ivan’s professional interests are in the areas of information retrieval, natural language processing, speech recognition and data mining.
Jason Dai is an Engineering Director and Principal Engineer in Intel SSG (Software and Services Group). He is currently leading the open source Hadoop engineering team in Intel; before that, he was the lead architect and engineering manager for building the 1st commercial auto-partitioning and parallelizing compiler on many-core many-thread processors (Intel Network Processor) in the industry.
Jimmy Lin is an Associate Professor at the University of Maryland who spent an extended sabbatical from 2010 to 2012 at Twitter, primarily working on relevance algorithms and analytics infrastructure.
Josh Patterson is a Principal Solution Architect at Cloudera. Prior to joining Cloudera, he was responsible for bringing Hadoop into the smartgrid during his involvement in the openPDC project. His focus in the smartgrid realm with Hadoop and HBase was using machine learning to discover and index anomalies in time series data. Josh is a graduate of the University of Tennessee at Chattanooga with a Bachelors in Business Management and a Masters of Computer Science with a thesis titled “TinyTermite: A Secure Routing Algorithm” where he worked in mesh networks and social insect swarm algorithms. Josh has over 15 years in software development and continues to contribute to projects such as Apach
Joydeep is a co-founder at Qubole and heads their India development team. Prior to starting Qubole, Joydeep worked at Facebook where he boot-strapped the data processing ecosystem based on Hadoop, started the Apache Hive project and led the Data Infrastructure team. Joydeep studied Computer Science at IIT-Delhi and University of Pittsburgh and has previously worked at Oracle, Netapp and Yahoo. Joydeep holds numerous patents, has many published papers and has been both speaker and panelist at Hadoop summits and at other Silicon Valley conferences.
As the Chief Technologist of T-Systems, Juergen Urbanski is the top executive responsible for the development and application of innovative technologies with the goal of growing revenue and profitability. As the technical lead on Deutsche Telekom’s four-member group-wide Cloud Leadership Team, he is accountable for driving the transformation of ICT services production towards cloud-enabled business outcomes such as rapid innovation, fast time to market, high capital efficiency and advanced cloud security. Previously, he was a Director of Product Management and Marketing with NetApp, where he had responsibility for growing the company’s Consulting and Integration business to over $60m in revenue. Prior, he spent 4 years with McKinsey and Company’s Silicon Valley office. He holds an MBA from The Wharton School.
Lars is the Director EMEA Services at Cloudera where he runs a team of Solutions Architects. He is involved with HBase and Hadoop since 2007 and a HBase committer since 2010. Lars is also the author of O`Reilly`s “HBase – The Definitive Guide”.
Mahadev is a core contributor and PMC member of Apache Ambari, Apache Hadoop and ZooKeeper. Prior to co-founding Hortonworks, Mahadev spent more than five years at Yahoo! working on Hadoop technologies. He holds a Master?s and Bachelor?s degree in Computer Science from SUNY Stony Brook and IIT Bombay (India) respectively.
Marilson Campos, has been involved in designing search engines and data pipelines since 1997. He has designed several big data architectures utilizing a diverse combination of tools and environments. He currently works for RocketFuel enabling the use of machine learning techniques on large datasets.
Martin Willcox’s career in IT spans 17 years, 5 organizations and 10 roles. In his current role (International Director of Technology Marketing), Martin is responsible for articulating Teradata’s strategy and the nature, value and differentiation of Teradata’s platform and solution offerings to International customers, prospects, partners,3rd party influencers and the media. Martin is a “poacher-turned-gamekeeper”, having been a Teradata customer for several years at Co-operative Retail before joining Teradata as an Enterprise Architecture consultant in Autumn 2004. Martin holds a BSc (Hons) in Physics and Astronomy from the University of Sheffield and a Postgraduate Certificate in Computing for Commerce and Industry from the Open University, is married with three children and is a lapsed supporter of Sheffield Wednesday Football Club. Martin is also an irregular contributor to the Teradata EMEA Blog and an occasional Twit (@willcoxmnk).
Matt is a Principal Program Manager on the Big Data team at Microsoft focused on building tools to make developers’ lives easier. Matt has been at Microsoft for 6 years working on developer tooling for distributed systems and cloud applications as well as technical evangelism.
Software Engineer at Cloudera, currently focused on the Apache HBase project.
Matthew Farrellee is a Principal Software Engineer at Red Hat and a researcher in the Computer Science department at the University of Wisconsin-Madison. Matthew has a rich development background in distributed computing, on which he writes a regular blog. Matthew has bachelor?s degrees in Computer Science and Cognitive Science from Indiana University and a master?s degree in Computer Science from the University of Wisconsin-Madison.
Michael Brown was a founding member of comScore, Inc. in 1999. He leads the company’s technology efforts to measure Internet and digital activities. He has been responsible for over 17 patent applications at comScore, five of which have already been issued by the US Patent and Trademark Office. Brown holds a Bachelor?s degree in Computer Science from the University of Maryland and a Masters in Computer and Information Science from Hood College.
Michael works at MapR EMEA in the role of Chief Data Engineer, where he helps people to tap the potential of big data. His background is in large-scale data integration research and development, advocacy and standardisation. He has experience with NoSQL databases and the Hadoop ecosystem. Michael contributes to Apache Drill, a distributed system for interactive analysis of large-scale datasets.
With extensive expertise in database systems from legacy to relational to traditional Data Warehouse to Big Data Mikhail Petrenko was able to successfully apply cutting edge data methodologies to deliver successful implementations in a wide variety of fields ? from digital marketing to medical research.
Neha is an Apache Kafka committer and member of the Apache Kafka PMC. Neha joined LinkedIn in early 2010 and has been working on the Distributed Data Systems team. She is one of the key developers of LinkedIn`s data pipeline and has been involved in all aspects of designing, developing and deploying Apache Kafka at LinkedIn. @nehanarkhede
Owen has been contributing to Apache Hadoop since before it was first called Hadoop. He was the first committer added to the project and has provided technical leadership on MapReduce, and security. Using Hadoop, in 2008 he set the world record for sorting a terabyte of data in 3.5 minutes and in 2009 he sorted a petabyte in 16.25 hours. In 2011, Owen co-founded Hortonworks, which commercially supports and trains users of the Hadoop ecosystem. Prior to Hortonworks, Owen worked on Yahoo! Search?s WebMap project, which built and performed heuristic analyses over a graph of the known web. Once ported to Apache Hadoop, it became the single largest known Hadoop application.
Patrick Angeles has over four years of experience in architecting, implementing and operationalizing Hadoop systems at various organizations.
Pramod Thangali has extensive experience building network, systems and cluster management products. Pramod joined Hortonworks in 2012 where he is leading the Apache Ambari initiative and related technologies to make it simple to install, provision and consume Hadoop in enterprise environments.
Raanan is responsible for Big Data Product Marketing at Splunk. He is focused on helping customers harness their big machine data to gain operational intelligence and driving an effective market-driven product strategy. Raanan has spent the past 10 years focused on data, information and distributed systems. Prior to Splunk, Raanan worked at Cloudera where he was a Hadoop sales engineer and before that, spent 9 years as an engineer and sales engineer at Oracle. Raanan is a certified Hadoop developer and administrator and holds a BS in Computer Science and MS in Management.
Robert is a Senior Software Engineer at Monsanto.
Rohit Bakshi works as product manager at Hortonworks
Sanjay is an Apache Hadoop committer and member of the Apache Hadoop PMC. Prior to co-founding Hortonworks, Sanjay was the chief architect of core-Hadoop at Yahoo and part of the team that created Hadoop. In Hadoop he has focused mostly on HDFS, MapReduce schedulers, high availability, compatibility, etc. He has also held senior engineering positions at Sun Microsystems and INRIA, where he developed software for distributed systems and grid/utility computing infrastructures. Sanjay has a PhD in Computer Science from the University of Waterloo in Canada.
Saurabh is employed currently in Hortonworks Professional Service group as Solution Architect. Saurabh has a total of 8.5 years of IT experience and worked in multiple assignments as a Big Data Architect, Performance Specialist and Database Administrator. He has provided and implemented Enterprise wide Big Data solution to solve existing bulk data processing and Analytic issues by Augmenting Data warehouse , Performance Engineering & Optimization for clients to improve overall system performance & response time with strong background in Hadoop ecosystem and Oracle database for varying clients of Energy, Retail and Financial Domains.
Siva Narayanan is works on Hadoop and Hive at Qubole. He’s previously worked on query optimization and workload management at EMC Greenplum. He has a Bachelor of Engineering from BITS, Pilani, India and a Ph.D. in Computer Science from The Ohio State University. Siva has several pending patents and has published many papers in academic conferences in the area of large-scale data management.
Steve Watt is a Consulting Software Engineer at Red Hat. Prior to that he spent 2 years as the Chief Technologist for Hadoop at Hewlett-Packard and 10 years working as a Researcher and Software Architect of Emerging Technologies within IBM Software Strategy and IBM Global Services. Steve is an Apache contributor, active in Open Source and chairs the Hadoop & Big Data User Group in Austin, Texas. Prior to working for HP and IBM, Steve spent several years consulting in the Middle East and working for startups in the United States and his native South Africa. Steve holds 22 US Patents and has published a number of technical books and articles.
Sujoe Bose is a Senior Principal at Sabre Holdings involved in the development of the Big Data Strategy for the Enterprise. He holds a PhD in Computer Science and Engineering from the University of Texas at Arlington and has more than 18 years of experience in Enterprise Systems design and development.
Suresh is an Apache Hadoop committer and member of the Apache Hadoop PMC. Suresh is an active contributor to the Apache HDFS project. Prior to co-founding Hortonworks, he served as a software architect at Yahoo! working on Apache Hadoop HDFS. Suresh also worked for Sylantro Systems in various technical leadership roles, developing scalable infrastructure for hosted communications services. He has a Bachelor?s degree in Electronics and Communication from the National Institute of Technology, Karnataka (India).
Sven Schlarb, PhD, studied Humanities Computer Science at the University of Cologne. He has many years of experience as a software developer in various companies and institutions in Cologne, Madrid, and Vienna. In 2008 he joined the Austrian National Library as a Software Developer with a focus on digitisation and long term preservation. In the past, he participated in the Europan projects PLANETS (www.planets-project.eu) and IMPACT (www.impact-project.eu). Now he is working in the SCAPE project (www.scape-project.eu) on the implementation of Hadoop-based solutions into the Austrian National Library`s IT infrastructure.
Ted has held Chief Scientist positions at Veoh Networks, ID Analytics and at MusicMatch, (now Yahoo Music). Ted is responsible for building the most advanced identity theft detection system on the planet, as well as one of the largest peer-assisted video distribution systems and ground-breaking music and video recommendations systems. Ted has 15 issued and 15 pending patents and contributes to several Apache open source projects including Hadoop, Zookeeper and Hbase?. He is also a committer for Apache Mahout and co-author of Mahout in Action.
Ulrich Rueckert is Data Scientist at Datameer. Prior to Datameer he worked as a research scholar at UC Berkeley and the International Computer Science Institute. His research on machine learning and data mining has been published in renowned journals and has won awards at international conferences. Ulrich serves on the program committees of the main machine learning conferences and he has organized workshops and held tutorials on his research.
Vin Sharma is a software strategist at Intel responsible for planning and marketing Intel contributions to open source datacenter software projects, specifically Hadoop, OpenStack, KVM, and enterprise Linux. In this role, Vin spearheads ecosystem initiatives with strategic partners and represents Intel on the board of the Open Virtualization Alliance and the OpenStack Foundation. Before joining Intel in 2011, Vin worked at HP for 15 years, most recently as the business strategist for open source and Linux on HP servers.
Vinod Kumar Vavilapalli is the Hadoop YARN and MapReduce guy at Hortonworks. He is a long term Hadoop contributor at Apache, Hadoop committer and a member of the Apache Hadoop PMC. He has a Bachelors degree from the Indian Institute of Technology Roorkee in Computer Science in Engineering. He has been working on Hadoop for more than 5 years and he still has fun doing it. He was involved in HadoopOnDemand, Hadoop-0.20, CapacityScheduler, Hadoop security, MapReduce and now is a lead developer for Hadoop YARN. He loves reading books, of all kinds. And passionate about using computers to change the world for better, bit by bit. He is reachable at twitter handle @tshooter.
Weiyan Wang is a software engineer at Data Infrastructure team at Facebook where he works on scaling Facebook’s warehouse cluster and using Erasure Codes in Hadoop distributed file system. Before Facebook, he interned at the stroage efficiency group at Netapp Inc. in 2010 and Microsoft Research Asia. in 2008. He received a MS in Computer Sciences from University of Wisconsin Madison and BS from Fudan University, China
William Heinzman is a software engineer at JNBridge (www.jnbridge.com). He has more than 25 years of experience in systems integration and developing manufacturing test, distributed job scheduling and IT software products. He currently develops interoperability solutions between Enterprise Java and .NET. He has a degree in Geophysics and a Masters of Engineering in Computer Science from the University of Colorado.