About The Position

Job Summary Responsible for collecting, storing, processing, and analyzing large sets of data. The primary focus will be on choosing optimal solutions to use for these purposes, then maintaining, implementing, and monitoring them. He/she will also be responsible to follow architecture and best practices used across the enterprise. Knowledge/Skills/Abilities: Define ideal Architecture, Evaluating tools and Frameworks, Standards & Best Practices for implementing scalable business solutions Implement Batch and Real-time data ingestion/extraction processes through ETL, Streaming, API, etc., between diverse source and target systems with structured and unstructured datasets Design and build data solutions with an emphasis on performance, scalability, and high-reliability Code, test, and document new or modified data systems to create robust and scalable applications for data analytics Build data model for analytics and application layers Working closely with multiple teams and Business partners, for collecting requirement and providing optimal solution Proven experience on Hadoop cluster components and services (like HDFS, YARN, ZOOKEEPER, AMBARI/CLOUDERA MANAGER, SENTRY/RANGER, KERBEROS, etc.) Ability to participate in troubleshooting technical issues while engaged with infrastructure and vendor support teams.

Requirements

  • 3 years of progressive experience using data, building ETLs (extract, transform, load), using disparate data sources for data models and analysis, and familiarity with data visualization tools and scripting tools such as Python, Ruby or Perl.
  • Experience in building stream-processing systems, using solutions such as Kafka, Storm or Spark-Streaming
  • Experience on Big Data tools such as, Spark, Hive, Impala, Polybase, Phoenix, Presto, Kylin, etc.
  • Experience with integration of data from multiple data sources (using ETL tool such, Talend, etc.)
  • Experience building solutions with NoSQL databases, such as HBase, Memsql
  • Strong experience on Database technologies, Data Warehouse, Data Validation & Certification, Data Quality, Metadata Management and Data Governance
  • Experience with programming language such as, Java/Scala/Python, etc.
  • Experience implementing Web application and Web Services APIs (REST/SOAP)

Nice To Haves

  • 5+ years of experience using data, building ETLs, using disparate data sources for data models and analysis and familiarity with data visualization tools and scripting tools such as Python, Ruby or Perl.
  • Experience in the healthcare industry

Responsibilities

  • Collecting, storing, processing, and analyzing large sets of data.
  • Choosing optimal solutions to use for data purposes, then maintaining, implementing, and monitoring them.
  • Following architecture and best practices used across the enterprise.
  • Defining ideal Architecture, Evaluating tools and Frameworks, Standards & Best Practices for implementing scalable business solutions
  • Implementing Batch and Real-time data ingestion/extraction processes through ETL, Streaming, API, etc., between diverse source and target systems with structured and unstructured datasets
  • Designing and building data solutions with an emphasis on performance, scalability, and high-reliability
  • Coding, testing, and documenting new or modified data systems to create robust and scalable applications for data analytics
  • Building data model for analytics and application layers
  • Working closely with multiple teams and Business partners, for collecting requirement and providing optimal solution
  • Participating in troubleshooting technical issues while engaged with infrastructure and vendor support teams.

Benefits

  • Molina Healthcare offers a competitive benefits and compensation package.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service