Sr. Big Data Architect - Spark (R1004299)

  • Company: IQVIA
  • Location: Plymouth Meeting, Pennsylvania
  • Posted: November 16, 2017
  • Reference ID: R1004299
Job Description

QuintilesIMS ( is the world's leading company providing information, technology, and services for virtually every type of stakeholder in healthcare. Every day, QuintilesIMS innovates on a large scale. We have been the big data company in healthcare before the "Big Data" term was defined. Since 1954 we have been helping healthcare organizations harness commercial and scientific insights and execute on their most strategic initiatives. We develop and operate systems and applications using big data technologies on our private-public hybrid cloud. QuintilesIMS stores and computes more than 20 petabytes of complex healthcare data on diseases, treatments, costs and outcomes to enable our clients to run their operations more efficiently. Our big data infrastructure is processing over 50 billion healthcare transactions annually. QuintilesIMS employs over 50,000 professionals in over 100 countries to drive results for over 5,000 healthcare clients globally. QuintilesIMS is a publicly listed company (NYSE - Q) with estimated revenue close to $7B and a market cap of over $18B.

We are looking for a Big Data Architect with solid Spark experience to validate the architecture and provide details and shared ownership for the architecture of our next generation data warehousing system, leading and working hands-on towards implementation and delivery to production
Help lead the charge on a data warehouse/operational data store strategy, ensuring rapid delivery while taking responsibility for applying standards, principles, theories, and concepts
Responsible for design and delivery of data models, which power BI initiatives, dashboards, syndicated reporting, and ad-hoc data exploratory canvases for IMS solutions
Work with data architects on the logical data models and physical database designs optimized for performance, availability and reliability
Tuning and optimization of backend and frontend data operations
Serve as a query tuning and optimization technical expert, providing feedback to team
Design and develop ETL and master data management processes
Scripting and automation to support development, QA and production database environments and deployments to production
Define and help enforce data governance and security policy
Mentors development team members
Proactively helps to resolve difficult technical issues
Provide technical knowledge to teams during project discovery and architecture phases
Keep management informed of work activities and schedules
Assess new initiatives to determine the work effort and estimate the necessary time-to-completion
Document new development, procedures or test plans as needed
Participate in data builds and deployment efforts
Participate in projects through various phases
Performs other related duties as assigned
Partner with the business units to develop effective solutions that solve business challenges

Required Experience (in order of importance):
Hadoop, Hive, Impala, and related technologies
Spark 1.6 / 2.0

MPP, shared nothing database systems, NoSQL systems
Mesos, Yarn

Lambda architectures
Data Warehousing design and concepts

Minimum Education, Experience, & Specialized Knowledge Required:
Computer Science Degree
5+ years strong native SQL skills
5+ years strong experience in database and data warehousing concepts and techniques. Must understand: relational and dimensional modeling, star/snowflake schema design, BI, Data Warehouse operating environments and related technologies, ETL, MDM, and data governance practices.
5+ years experience working in Linux
3+ years experience with Hadoop, Hive, Impala, HBase, and related technologies
2+ years strong experience with low latency (near real time) systems and working with Tb data sets, loading and processing billions of records per day
1+ years' experience with MapReduce/YARN
2+ years' experience with Spark
1+ years experience with Lambda architectures
1+ years experience with MPP, shared nothing database systems, and NoSQL systems
Ability to work in a fast-paced, team-oriented environment
Ability to complete the full lifecycle of software development and deliver on time
Ability to work with end-users to gather requirements and convert them to working documents
Strong interpersonal skills, including a positive, solution-oriented attitude
Must be passionate, flexible and innovative in utilizing the tools, their experience, and any other resources, to effectively deliver to very challenging and always changing business requirements with continuous success
Must be able to interface with various solution/business areas to understand the requirements and prepare documentation to support development
Healthcare and/or reference data experience is a plus



Share this Job