Big Data Architect

About Capgemini <?xml:namespace prefix = o ns = "urn:schemas-microsoft-com:office:office" />

With more than 170,000 people in over 40 countries, Capgemini is one of the world's foremost providers of consulting, technology and outsourcing services. The Group reported 2014 global revenues of EUR 10.5 billion. Together with its clients, Capgemini creates and delivers business and technology solutions that fit their needs and drive the results they want. A deeply multicultural organization, Capgemini has developed its own way of working, the Collaborative Business ExperienceTM, and draws on Rightshore®, its worldwide delivery model.

Learn more about us at http://www.capgemini.com/ .

 

Rightshore ® is a trademark belonging to Capgemini

Capgemini is an Equal Opportunity Employer encouraging diversity in the workplace. All qualified applicants will receive consideration for employment without regard to race, national origin, gender identity/expression, age, religion, disability, sexual orientation, genetics, veteran status, marital status or any other characteristic protected by law.

 

Title: Big Data Architect

 

Qualifications:

10+ years of solid IT consulting experience in data warehousing, VLDBs, operational data stores and large scale implementations

Experience working as a member of a distributed, often global, team of technical and domain experts using agile development methods

Excellent one-on-one communication and presentation skills, able to convey technical information in a clear and unambiguous manner. 

 

Role/Responsibilities:

Design and build scalable infrastructure and Big Data platforms to collect and process very large amounts of

Data (structured and unstructured), including if applicable streaming real-time data.

 

Work closely across an array of various teams and organizations in the company and industry

(Including partners, customers and researchers) to:

 

Demonstrate thought leadership and guide customers on Big Data adoption, setting up best practices, governance structures

Act as internal resource to help teams leverage our big data capabilities

Effectively communicate solutions architecture to key stakeholders and project teams

Evaluate and adopt latest technologies to develop cloud based big data and real-time analytics solutions

Contribute on multiple Big Data projects and assign tasks to junior engineers, oversee the execution of tasks and provide mentorship and guidance as needed

Understand business objectives and suggests technical strategies to meet those objectives

Contribute to business requirement definition and use case design as a technical expert

Convert business requirements into architectural designs and detailed technical designs

Identify tasks, effort and dependencies based on software architecture and specifications

Guide performance testing and recommend solutions for any performance bottle necks

Plan and execute technology proof-of-concept’s (POC) using Big Data technology

Undertake feasibility studies for major IT developments incorporating costs and benefits, and presents proposals to clients;

Examine existing business models and flows of data and designs functional specifications and test plans for new systems in order to meet clients needs;

 

Using Big Data technologies and given business requirements, design a comprehensive technical architecture including:

Conceptual, Logical, Physical Models

Data quality standards

Data indexing

Extraction and analytic queries

 

Skills (mandatory):

Problem solver by nature with a strong sense of curiosity about technical puzzles especially large scale distributed computing 

 

Hands on technical competence in one or more of following:

Programming languages - Java/J2EE, Linux, PHP, Python, C, C++, Hadoop, Hive, HBase, Pig, MapReduce and other Hadoop eco-system components, R on Hadoop, Mahout

Data warehouse, BI and ETL tools

Detailed knowledge of RDBMS data modeling and SQL

Some knowledge of NoSQL databases types such as OLAP, Graph, Key Value, Object, XML, JSON, SQL, NOSQL, Tuple Store, Columnar, in-memory

 

Skills (nice to have): 

Other relevant Apache technologies including:

Storm, Kafka, Solr/Lucene,

Enterprise Application integration

Master Data Management tools and methodologies 

 

Location: New York, NY OR Rosemont, IL