OnlyDataJobs.com

Acxiom
  • Austin, TX
As a Senior Hadoop Administrator, you will assist leadership for projects related to Big Data technologies and software development support for client research projects. You will analyze the latest Big Data Analytic technologies and their innovative applications in both business intelligence analysis and new service offerings. You will bring these insights and best practices to Acxiom's Big Data Projects. You are able to benchmark systems, analyze system bottlenecks and propose solutions to eliminate them. You will develop highly scalable and extensible Big Data platform which enables collection, storage, modeling, and analysis of massive data sets from numerous channels. You are also a self-starter able to continuously evaluate new technologies, innovate and deliver solutions for business critical applications


What you will do:


  • Responsible for implementation and ongoing administration of Hadoop infrastructure
  • Provide technical leadership and collaboration with engineering organization, develop key deliverables for Data Platform Strategy - Scalability, optimization, operations, availability, roadmap.
  • Lead the platform architecture and drive it to the next level of effectiveness to support current and future requirements
  • Cluster maintenance as well as creation and removal of nodes using tools like Cloudera Manager Enterprise, etc.
  • Performance tuning of Hadoop clusters and Hadoop MapReduce routines
  • Screen Hadoop cluster job performances and capacity planning
  • Help optimize and integrate new infrastructure via continuous integration methodologies (DevOps CHEF)
  • Lead and review Hadoop log files with the help of log management technologies (ELK)
  • Provide top-level technical help desk support for the application developers
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality, availability and security
  • Collaborating with application teams to perform Hadoop updates, patches, version upgrades when required
  • Mentor Hadoop engineers and administrators
  • Work with vendor support teams on support tasks


Do you have?


  • Bachelor's degree in related field of study, or equivalent experience
  • 6+ years of Big Data Administration Experience
  • Extensive knowledge and Hands-on Experience of Hadoop based data manipulation/storage technologies like HDFS, MapReduce, Yarn, Spark/Kafka, HBASE, HIVE, Pig, Impala, R and Sentry/Ranger/Knox
  • Experience in capacity planning, cluster designing and deployment, troubleshooting and performance tuning
  • Experience supporting Data Science teams and Analytics teams on complex code deployment, debugging and performance optimization problems
  • Great operational expertise such as excellent troubleshooting skills, understanding of system's capacity, bottlenecks, core resource utilizations (CPU, OS, Storage, and Networks)
  • Experience in Hadoop cluster migrations or upgrades
  • Strong scripting skills in Perl, Python, shell scripting, and/or Ruby on Rails
  • Linux/SAN administration skills and RDBMS/ETL knowledge
  • Good Experience in Cloudera, HortonWorks, and/or MapR versions along with monitoring/alerting tools (Nagios, Ganglia, Zenoss, Cloudera Manager)
  • Strong problem solving and critical thinking skills
  • Excellent verbal and written communication skills


What will set you apart:


  • Solid understanding and hands-on experience of Big Data on private/public cloud technologies(AWS/GCP/Azure)
  • DevOps experience (CHEF, Puppet and Ansible)
  • Strong knowledge of JAVA/J2EE and other web technologies

 
TalentBurst, an Inc 5000 company
  • Austin, TX

Position: Hadoop Administrator

Location: Austin, TX

Duration: 12+ months

Interview: Skype

Job Description:
Person will be responsible to work as part of 24x7 shifts (US hours) to provide Hadoop Platform Support and Perform Adminstrative on Production Hadoop clusters.

Must have skills:
3+ yrs. hands-on administration experience on Large Distributed Hadoop System on Linux

Technical Knowledge on YARN, MapReduce, HDFS, HBase, Zookeeper, Pig and Hive

Hands-on Experience as a Linux Sys Admin

Nice to have skills:
Knowledge on Spark and Kafka is a plus / Hadoop Certification is preferred

Roles and responsibilities:
Hadoop cluster set up, performance fine-tuning, monitoring and administration

Skill Requirements:
Minimum 3 yrs. hands-on experience on Large Distributed Hadoop System on Linux.
Strong Technical Knowledge on Hadoop Eco System such as YARN, MapReduce, HDFS, HBase, Zookeeper, Pig and Hive.
Hands on experience in Hadoop cluster set up, performance fine-tuning, monitoring and administration.
Hands-on Experience as a Linux Sys Admin
Knowledge on Spark and Kafka is a plus.
Hadoop Certification is preferred

Comcast
  • Philadelphia, PA

Comcast's Technology & Product organization works at the intersection of media and technology. Our innovative teams are continually developing and delivering products that transform the customer experience. From creating apps like TVGo to new features such as the Talking Guide on the X1 platform, we work every day to make a positive impact through innovation in the pursuit of building amazing products that are enjoyable, easy to use and accessible across all platforms. The team also develops and supports our evolving network architecture, including next-generation consumer systems and technologies, infrastructure and engineering, network integration and management tools, and technical standards.

As a Data Science Engineer in Comcastdx, you will research, model, develop, support data pipelines and deliver insights for key strategic initiatives. You will develop or utilize complex programmatic and quantitative methods to find patterns and relationships in data sets; lead statistical modeling, or other data-driven problem-solving analysis to address novel or abstract business operation questions; and incorporate insights and findings into a range of products.

Assist in design and development of collection and enrichment components focused on quality, timeliness, scale and reliability. Work on real-time data stores and a massive historical data store using best-of-breed and industry leading technology.

Responsibilities:

-Develop and support data pipelines

-Analyze massive amounts of data both in real-time and batch processing utilizing Spark, Kafka, and AWS technologies such as Kinesis, S3, ElastiSearch, and Lambda

-Create detailed write-ups of processes used, logic applied, and methodologies used for creation, validation, analysis, and visualizations. Write ups shall occur initially, within a week of when process is created, and updated in writing when changes occur.

-Prototype ideas for new ML/AI tools, products and services

-Centralize data collection and synthesis, including survey data, enabling strategic and predictive analytics to guide business decisions

-Provide expert and professional data analysis to implement effective and innovative solutions meshing disparate data types to discover insights and trends.

-Employ rigorous continuous delivery practices managed under an agile software development approach

-Support DevOps practices to deploy and operate our systems

-Automate and streamline our operations and processes

-Troubleshoot and resolve issues in our development, test and production environments

Here are some of the specific technologies and concepts we use:

-Spark Core and Spark Streaming

-Machine learning techniques and algorithms

-Java, Scala, Python, R

-Artificial Intelligence

-AWS services including EMR, S3, Lambda, ElasticSearch

-Predictive Analytics

-Tableau, Kibana

-Git, Maven, Jenkins

-Linux

-Kafka

-Hadoop (HDFS, YARN)

Skills & Requirements:

-5-8 years of Java experience, Scala and Python experience a plus

-3+ years of experience as an analyst, data scientist, or related quantitative role.

-3+ years of relevant quantitative and qualitative research and analytics experience. Solid knowledge of statistical techniques.

-Bachelors in Statistics, Math, Engineering, Computer Science, Statistics or related discipline. Master's Degree preferred.

-Experience in software development of large-scale distributed systems including proven track record of delivering backend systems that participate in a complex ecosystem

-Experience with more advanced modeling techniques (eg ML.)

-Distinctive problem solving and analysis skills and impeccable business judgement.

-Experience working with imperfect data sets that, at times, will require improvements to process, definition and collection

-Experience with real-time data pipelines and components including Kafka, Spark Streaming

-Proficient in Unix/Linux environments

-Test-driven development/test automation, continuous integration, and deployment automation

-Excellent communicator, able to analyze and clearly articulate complex issues and technologies understandably and engagingly

-Team player is a must

-Great design and problem-solving skills

-Adaptable, proactive and willing to take ownership

-Attention to detail and high level of commitment

-Thrives in a fast-paced agile environment

About Comcastdx:

Comcastdxis a result driven big data engineering team responsible for delivery of multi-tenant data infrastructure and platforms necessary to support our data-driven culture and organization.dxhas an overarching objective to gather, organize, and make sense of Comcast data with intention to reveal business and operational insight, discover actionable intelligence, enable experimentation, empower users, and delight our stakeholders. Members of thedxteam define and leverage industry best practices, work on large-scale data problems, design and develop resilient and highly robust distributed data organizing and processing systems and pipelines as well as research, engineer, and apply data science and machine intelligence disciplines.

Comcast is an EOE/Veterans/Disabled/LGBT employer

Comcast
  • Philadelphia, PA

Comcast's Technology & Product organization works at the intersection of media and technology. Our innovative teams are continually developing and delivering products that transform the customer experience. From creating apps like TVGo to new features such as the Talking Guide on the X1 platform, we work every day to make a positive impact through innovation in the pursuit of building amazing products that are enjoyable, easy to use and accessible across all platforms. The team also develops and supports our evolving network architecture, including next-generation consumer systems and technologies, infrastructure and engineering, network integration and management tools, and technical standards.

As a Data Science Engineer in Comcastdx, you will research, model, develop, support data pipelines and deliver insights for key strategic initiatives. You will develop or utilize complex programmatic and quantitative methods to find patterns and relationships in data sets; lead statistical modeling, or other data-driven problem-solving analysis to address novel or abstract business operation questions; and incorporate insights and findings into a range of products.

Assist in design and development of collection and enrichment components focused on quality, timeliness, scale and reliability. Work on real-time data stores and a massive historical data store using best-of-breed and industry leading technology.

Responsibilities:

-Develop and support data pipelines

-Analyze massive amounts of data both in real-time and batch processing utilizing Spark, Kafka, and AWS technologies such as Kinesis, S3, ElastiSearch, and Lambda

-Create detailed write-ups of processes used, logic applied, and methodologies used for creation, validation, analysis, and visualizations. Write ups shall occur initially, within a week of when process is created, and updated in writing when changes occur.

-Prototype ideas for new ML/AI tools, products and services

-Centralize data collection and synthesis, including survey data, enabling strategic and predictive analytics to guide business decisions

-Provide expert and professional data analysis to implement effective and innovative solutions meshing disparate data types to discover insights and trends.

-Employ rigorous continuous delivery practices managed under an agile software development approach

-Support DevOps practices to deploy and operate our systems

-Automate and streamline our operations and processes

-Troubleshoot and resolve issues in our development, test and production environments

Here are some of the specific technologies and concepts we use:

-Spark Core and Spark Streaming

-Machine learning techniques and algorithms

-Java, Scala, Python, R

-Artificial Intelligence

-AWS services including EMR, S3, Lambda, ElasticSearch

-Predictive Analytics

-Tableau, Kibana

-Git, Maven, Jenkins

-Linux

-Kafka

-Hadoop (HDFS, YARN)

Skills & Requirements:

-3-5years of Java experience, Scala and Python experience a plus

-2+ years of experience as an analyst, data scientist, or related quantitative role.

-2+ years of relevant quantitative and qualitative research and analytics experience. Solid knowledge of statistical techniques.

-Bachelors in Statistics, Math, Engineering, Computer Science, Statistics or related discipline. Master's Degree preferred.

-Experience in software development of large-scale distributed systems including proven track record of delivering backend systems that participate in a complex ecosystem

-Experience with more advanced modeling techniques (eg ML.)

-Distinctive problem solving and analysis skills and impeccable business judgement.

-Experience working with imperfect data sets that, at times, will require improvements to process, definition and collection

-Experience with real-time data pipelines and components including Kafka, Spark Streaming

-Proficient in Unix/Linux environments

-Test-driven development/test automation, continuous integration, and deployment automation

-Excellent communicator, able to analyze and clearly articulate complex issues and technologies understandably and engagingly

-Team player is a must

-Great design and problem-solving skills

-Adaptable, proactive and willing to take ownership

-Attention to detail and high level of commitment

-Thrives in a fast-paced agile environment

About Comcastdx:

Comcastdxis a result driven big data engineering team responsible for delivery of multi-tenant data infrastructure and platforms necessary to support our data-driven culture and organization.dxhas an overarching objective to gather, organize, and make sense of Comcast data with intention to reveal business and operational insight, discover actionable intelligence, enable experimentation, empower users, and delight our stakeholders. Members of thedxteam define and leverage industry best practices, work on large-scale data problems, design and develop resilient and highly robust distributed data organizing and processing systems and pipelines as well as research, engineer, and apply data science and machine intelligence disciplines.

Comcast is an EOE/Veterans/Disabled/LGBT employer

Visa
  • Austin, TX
Company Description
Common Purpose, Uncommon
Opportunity. Everyone at Visa works with one goal in mind making sure that Visa is the best way to pay and be paid, for everyone everywhere. This is our global vision and the common purpose that unites the entire Visa team. As a global payments technology company, tech is at the heart of what we do: Our VisaNet network processes over 13,000 transactions per second for people and businesses around the world, enabling them to use digital currency instead of cash and checks. We are also global advocates for financial inclusion, working with partners around the world to help those who lack access to financial services join the global economy. Visas sponsorships, including the Olympics and FIFA World Cup, celebrate teamwork, diversity, and excellence throughout the world. If you have a passion to make a difference in the lives of people around the
world, Visa offers an uncommon opportunity to build a strong, thriving career. Visa is fueled by our team of talented employees who continuously raise the bar on delivering the convenience and security of digital currency to people all over the world. Join our team and find out how Visa is everywhere you want to
be.
Job Description
The ideal candidate will be responsible for the following to:
  • Perform Hadoop Administration on Production Hadoop clusters
  • Perform Tuning and Increase Operational efficiency on a continuous basis
  • Monitor health of the platforms and Generate Performance Reports and Monitor and provide continuous improvements
  • Working closely with development, engineering and operation teams, jointly work on key deliverables ensuring production scalability and stability
  • Develop and enhance platform best practices
  • Ensure the Hadoop platform can effectively meet performance & SLA requirements
  • Responsible for support of Hadoop Production environment which includes Hive, YARN, Spark, Impala, Kafka, SOLR, Oozie, Sentry, Encryption, Hbase, etc.
  • Perform optimization capacity planning of a large multi-tenant cluster
Qualifications
  • Minimum 3 years of work experience in maintaining, optimization, issue resolution of Hadoop clusters, supporting Business users and Batch
  • Experience in Configuring and setting up Hadoop clusters and provide support for - aggregation, lookup & fact table creation criteria
  • Map Reduce tuning, data node, NN recovery etc.
  • Experience in Linux / Unix OS Services, Administration, Shell, awk scripting
  • Experience in building and scalable Hadoop applications
  • Experience in Core Java, Hadoop (Map Reduce, Hive, Pig, HDFS, H-catalog, Zookeeper and OOzie)
  • Hands-on Experience in SQL (Oracle ) and No SQL Databases (HBASE/Cassandra/Mongo DB)
  • Excellent oral and written communication and presentation skills, analytical and problem solving skills
  • Self-driven, Ability to work independently and as part of a team with proven track record developing and launching products at scale
  • Minimum of four year technical degree required
  • Experience on Cloudera distribution preferred
  • Hands-on Experience as a Linux Sys Admin is a plus
  • Knowledge on Spark and Kafka is a plus.
Additional Information
All your information will be kept confidential according to EEO guidelines.
Job Number: REF15232V
phData, Inc.
  • Minneapolis, MN

Title: Big Data Solutions Architect (Minneapolis or US Remote)


Join the Game-Changers in Big Data  


Are you inspired by innovation, hard work and a passion for data?    


If so, this may be the ideal opportunity to leverage your background in Big Data and Software Engineering, Data Engineering or Data Analytics experience to design, develop and innovate big data solutions for a diverse set of clients.  


As a Solution Architect on our Big Data Consulting team, your responsibilities include:


    • Design, develop, and innovative Big Data solutions; partner with our internal Managed Services Architects and Data Engineers to build creative solutions to solve tough big data problems.  
    • Determine the project road map, select the best tools, assign tasks and priorities, and assume general project management oversight for performance, data integration, ecosystem integration, and security of big data solutions
    • Work across a broad range of technologies from infrastructure to applications to ensure the ideal Big Data solution is implemented and optimized
    • Integrate data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures (AWS); determine new and existing data sources
    • Design and implement streaming, data lake, and analytics big data solutions

    • Create and direct testing strategies including unit, integration, and full end-to-end tests of data pipelines

    • Select the right storage solution for a project - comparing Kudu, HBase, HDFS, and relational databases based on their strengths

    • Utilize ETL processes to build data repositories; integrate data into Hadoop data lake using Sqoop (batch ingest), Kafka (streaming), Spark, Hive or Impala (transformation)

    • Partner with our Managed Services team to design and install on prem or cloud based infrastructure including networking, virtual machines, containers, and software

    • Determine and select best tools to ensure optimized data performance; perform Data Analysis utilizing Spark, Hive, and Impala

    • Mentor and coach Developers and Data Engineers. Provide guidance with project creation, application structure, automation, code style, testing, and code reviews

Qualifications

  • 5+ years previous experience as a Software Engineer, Data Engineer or Data Analytics - combined with an expertise in Hadoop Technologies and Java programming
  • Technical Leadership experience leading/mentoring junior software/data engineers, as well as scoping activities on large scale, complex technology projects
  • Expertise in core Hadoop technologies including HDFS, Hive and YARN.  
  • Deep experience in one or more ecosystem products/languages such as HBase, Spark, Impala, Solr, Kudu, etc
  • Expert programming experience in Java, Scala, or other statically typed programming language
  • Strong working knowledge of SQL and the ability to write, debug, and optimize distributed SQL queries
  • Excellent communication skills including proven experience working with key stakeholders and customers
  • Ability to translate big picture business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access and consumption, as well as custom analytics
  • Customer relationship management including project escalations, and participating in executive steering meetings
  • Ability to learn new technologies in a quickly changing field
phData, Inc.
  • Minneapolis, MN

Title: Big Data Solutions Architect (Minneapolis or US Remote)


Join the Game-Changers in Big Data  


Are you inspired by innovation, hard work and a passion for data?    


If so, this may be the ideal opportunity to leverage your background in Big Data and Software Engineering, Data Engineering or Data Analytics experience to design, develop and innovate big data solutions for a diverse set of clients.  


As a Solution Architect on our Big Data Consulting team, your responsibilities include:


    • Design, develop, and innovative Big Data solutions; partner with our internal Managed Services Architects and Data Engineers to build creative solutions to solve tough big data problems.  
    • Determine the project road map, select the best tools, assign tasks and priorities, and assume general project management oversight for performance, data integration, ecosystem integration, and security of big data solutions
    • Work across a broad range of technologies from infrastructure to applications to ensure the ideal Big Data solution is implemented and optimized
    • Integrate data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures (AWS); determine new and existing data sources
    • Design and implement streaming, data lake, and analytics big data solutions

    • Create and direct testing strategies including unit, integration, and full end-to-end tests of data pipelines

    • Select the right storage solution for a project - comparing Kudu, HBase, HDFS, and relational databases based on their strengths

    • Utilize ETL processes to build data repositories; integrate data into Hadoop data lake using Sqoop (batch ingest), Kafka (streaming), Spark, Hive or Impala (transformation)

    • Partner with our Managed Services team to design and install on prem or cloud based infrastructure including networking, virtual machines, containers, and software

    • Determine and select best tools to ensure optimized data performance; perform Data Analysis utilizing Spark, Hive, and Impala

    • Mentor and coach Developers and Data Engineers. Provide guidance with project creation, application structure, automation, code style, testing, and code reviews

Qualifications

  • 5+ years previous experience as a Software Engineer, Data Engineer or Data Analytics - combined with an expertise in Hadoop Technologies and Java programming
  • Technical Leadership experience leading/mentoring junior software/data engineers, as well as scoping activities on large scale, complex technology projects
  • Expertise in core Hadoop technologies including HDFS, Hive and YARN.  
  • Deep experience in one or more ecosystem products/languages such as HBase, Spark, Impala, Solr, Kudu, etc
  • Expert programming experience in Java, Scala, or other statically typed programming language
  • Strong working knowledge of SQL and the ability to write, debug, and optimize distributed SQL queries
  • Excellent communication skills including proven experience working with key stakeholders and customers
  • Ability to translate big picture business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access and consumption, as well as custom analytics
  • Customer relationship management including project escalations, and participating in executive steering meetings
  • Ability to learn new technologies in a quickly changing field
Acxiom
  • Austin, TX
As a Hadoop Administrator, you will assist leadership for projects related to Big Data technologies and software development support for client research projects. You will analyze latest Big Data Analytic technologies and their innovative applications in both business intelligence analysis and new service offerings. You will bring these insights and best practices to Acxiom's Big Data Projects. You must be able to benchmark systems, analyze system bottlenecks and propose solutions to eliminate them. You will develop highly scalable and extensible Big Data platform which enables collection, storage, modeling, and analysis of massive data sets from numerous channels. You must be a self-starter to continuously evaluate new technologies, innovate and deliver solutions for business critical applications. 


 

What you will do:


  • Responsible for implementation and ongoing administration of Hadoop infrastructure
  • Provide technical leadership and collaboration with engineering organization, develop key deliverables for Data Platform Strategy - Scalability, optimization, operations, availability, roadmap.
  • Own the platform architecture and drive it to the next level of effectiveness to support current and future requirements
  • Cluster maintenance as well as creation and removal of nodes using tools like Cloudera Manager Enterprise, etc.
  • Performance tuning of Hadoop clusters and Hadoop MapReduce routines
  • Screen Hadoop cluster job performances and capacity planning
  • Help optimize and integrate new infrastructure via continuous integration methodologies (DevOps CHEF)
  • Manage and review Hadoop log files with the help of  Log management technologies (ELK)
  • Provide top-level technical help desk support for the application developers
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality, availability and security
  • Collaborating with application teams to perform Hadoop updates, patches, version upgrades when required
  • Mentor Hadoop engineers and administrators
  • Work with Vendor support teams on support tasks


Do you have?


  • Bachelor's degree in related field of study, or equivalent experience
  • 3+ years of Big Data Administration experience
  • Extensive knowledge of Hadoop based data manipulation/storage technologies such as HDFS, MapReduce, Yarn, HBASE, HIVE, Pig, Impala and Sentry
  • Experience in capacity planning, cluster designing and deployment, troubleshooting and performance tuning
  • Great operational expertise such as good troubleshooting skills, understanding of system's capacity, bottlenecks, basics of memory, CPU, OS, storage, and networks
  • Experience in Hadoop cluster migrations or upgrades
  • Strong Linux/SAN administration skills and RDBMS/ETL knowledge
  • Good Experience in Cloudera/Horton Works/MapR versions along with Monitoring/Alerting tools (Nagios, Ganglia, Zenoss , Cloudera Manager)
  • Scripting skills in Perl, Python, Shell Scripting, and/or Ruby on Rails
  • Knowledge of JAVA/J2EE and other web technologies
  • Understanding of On-premise and Cloud network architectures
  • DevOps experience is a great plus (CHEF, Puppet and Ansible)
  • Excellent verbal and written communication skills


 

Perficient, Inc.
  • Dallas, TX
At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
Perficient currently has a career opportunity for a Big Data Engineer(Microservices Developer),
Job Overview
One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics. As a lead developer, you will be responsible for Microservices development.
Responsibilities
  • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFSWorked with different data formats (Parquet, Avro, JSON, XML, etc.)
  • Worked on containerized solutions (Kubernetes..)
  • Provide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE
  • Should be a Rockstar in HBase and Hive Best Practices
  • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFS
  • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
  • Worked on containerized solutions (Spring Boot and Docker)
  • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
  • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
  • Performance tune HIVE and HBASE jobs with a focus on ingestion
  • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
  • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
  • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
  • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
Qualifications
    • Spring, Docker, Hibernate /Spring , JPA, Pivotal, Kafka, NoSQL,
      Hadoop Containers Docker, work, Spring boot .
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolsets
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • Houston, TX
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a MapR Spark Developer
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics.
    As a Data developer, you will be incharge of developing data pipelines using Spark and other ingestion framework.
    Responsibilities
      Prov
      • ide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE Shoul
      • d be a Rockstar in HBase and Hive Best PracticesAbili
      • ty to focus on ingestion and transformationUnder
      • stand MDM and think about MDM in HDFSAbili
      • ty to design solutions for different use cases
      • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
      • Troubleshoot and develop on Hadoop technologies including HDFS, Kafka, Hive, Pig, Flume, HBase, Spark, Impala and Hadoop ETL development via tools such as ODI for Big Data and APIs to extract data from source.
      • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
      • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
      • Performance tune HIVE and HBASE jobs with a focus on ingestion
      • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
      • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
      • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
      • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolset
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • Salt Lake City, UT
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a Big Data Engineer(Microservices Developer),
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics. As a lead developer, you will be responsible for Microservices development.
    Responsibilities
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFSWorked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Kubernetes..)
    • Provide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE
    • Should be a Rockstar in HBase and Hive Best Practices
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFS
    • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Spring Boot and Docker)
    • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
    • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
    • Performance tune HIVE and HBASE jobs with a focus on ingestion
    • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
    • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
    • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
    • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • Spring, Docker, Hibernate /Spring , JPA, Pivotal, Kafka, NoSQL,
      Hadoop Containers Docker, work, Spring boot .
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolsets
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • Philadelphia, PA
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a Big Data Engineer(Microservices Developer),
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics. As a lead developer, you will be responsible for Microservices development.
    Responsibilities
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFSWorked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Kubernetes..)
    • Provide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE
    • Should be a Rockstar in HBase and Hive Best Practices
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFS
    • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Spring Boot and Docker)
    • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
    • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
    • Performance tune HIVE and HBASE jobs with a focus on ingestion
    • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
    • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
    • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
    • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • Spring, Docker, Hibernate /Spring , JPA, Pivotal, Kafka, NoSQL,
      Hadoop Containers Docker, work, Spring boot .
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolsets
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • Atlanta, GA
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a MapR Spark Developer
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics.
    As a Data developer, you will be incharge of developing data pipelines using Spark and other ingestion framework.
    Responsibilities
      Prov
      • ide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE Shoul
      • d be a Rockstar in HBase and Hive Best PracticesAbili
      • ty to focus on ingestion and transformationUnder
      • stand MDM and think about MDM in HDFSAbili
      • ty to design solutions for different use cases
      • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
      • Troubleshoot and develop on Hadoop technologies including HDFS, Kafka, Hive, Pig, Flume, HBase, Spark, Impala and Hadoop ETL development via tools such as ODI for Big Data and APIs to extract data from source.
      • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
      • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
      • Performance tune HIVE and HBASE jobs with a focus on ingestion
      • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
      • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
      • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
      • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolset
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • Ann Arbor, MI
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a MapR Spark Developer
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics.
    As a Data developer, you will be incharge of developing data pipelines using Spark and other ingestion framework.
    Responsibilities
      Prov
      • ide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE Shoul
      • d be a Rockstar in HBase and Hive Best PracticesAbili
      • ty to focus on ingestion and transformationUnder
      • stand MDM and think about MDM in HDFSAbili
      • ty to design solutions for different use cases
      • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
      • Troubleshoot and develop on Hadoop technologies including HDFS, Kafka, Hive, Pig, Flume, HBase, Spark, Impala and Hadoop ETL development via tools such as ODI for Big Data and APIs to extract data from source.
      • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
      • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
      • Performance tune HIVE and HBASE jobs with a focus on ingestion
      • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
      • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
      • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
      • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolset
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • Minneapolis, MN
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a Big Data Engineer(Microservices Developer),
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics. As a lead developer, you will be responsible for Microservices development.
    Responsibilities
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFSWorked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Kubernetes..)
    • Provide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE
    • Should be a Rockstar in HBase and Hive Best Practices
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFS
    • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Spring Boot and Docker)
    • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
    • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
    • Performance tune HIVE and HBASE jobs with a focus on ingestion
    • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
    • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
    • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
    • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • Spring, Docker, Hibernate /Spring , JPA, Pivotal, Kafka, NoSQL,
      Hadoop Containers Docker, work, Spring boot .
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolsets
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • San Diego, CA
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a MapR Spark Developer
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics.
    As a Data developer, you will be incharge of developing data pipelines using Spark and other ingestion framework.
    Responsibilities
      Prov
      • ide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE Shoul
      • d be a Rockstar in HBase and Hive Best PracticesAbili
      • ty to focus on ingestion and transformationUnder
      • stand MDM and think about MDM in HDFSAbili
      • ty to design solutions for different use cases
      • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
      • Troubleshoot and develop on Hadoop technologies including HDFS, Kafka, Hive, Pig, Flume, HBase, Spark, Impala and Hadoop ETL development via tools such as ODI for Big Data and APIs to extract data from source.
      • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
      • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
      • Performance tune HIVE and HBASE jobs with a focus on ingestion
      • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
      • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
      • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
      • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolset
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • San Diego, CA
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a Big Data Engineer(Microservices Developer),
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics. As a lead developer, you will be responsible for Microservices development.
    Responsibilities
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFSWorked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Kubernetes..)
    • Provide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE
    • Should be a Rockstar in HBase and Hive Best Practices
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFS
    • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Spring Boot and Docker)
    • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
    • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
    • Performance tune HIVE and HBASE jobs with a focus on ingestion
    • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
    • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
    • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
    • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • Spring, Docker, Hibernate /Spring , JPA, Pivotal, Kafka, NoSQL,
      Hadoop Containers Docker, work, Spring boot .
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolsets
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?
    Perficient, Inc.
    • Austin, TX
    At Perficient youll deliver mission-critical technology and business solutions to Fortune 500 companies and some of the most recognized brands on the planet. And youll do it with cutting-edge technologies, thanks to our close partnerships with the worlds biggest vendors. Our network of offices across North America, as well as locations in India and China, will give you the opportunity to spread your wings, too.
    Were proud to be publicly recognized as a Top Workplace year after year. This is due, in no small part, to our entrepreneurial attitude and collaborative spirit that sets us apart and keeps our colleagues impassioned, driven, and fulfilled.
    Perficient currently has a career opportunity for a Big Data Engineer(Microservices Developer),
    Job Overview
    One of our large clients has made strategic decision to move all order management and sales data from their existing EDW into MapR platform. The focus is fast ingestion and streaming analytics. This is a multiyear roadmap with many components that will piece into a larger Data Management Platform. Perficient subject matter expert will work with the client team to move this data into new environment in a fashion that will meet requirements for applications and analytics. As a lead developer, you will be responsible for Microservices development.
    Responsibilities
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFSWorked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Kubernetes..)
    • Provide end to end vision and hands on experience with MapR Platform especially best practices around HIVE and HBASE
    • Should be a Rockstar in HBase and Hive Best Practices
    • Ability to focus on framework for Dev Ops, Ingestion, and Reading/writing into HDFS
    • Worked with different data formats (Parquet, Avro, JSON, XML, etc.)
    • Worked on containerized solutions (Spring Boot and Docker)
    • Translate, load and present disparate data-sets in multiple formats and from multiple sources including JSON, Avro, text files, Kafka queues, and log data.
    • Lead workshops with many teams to define data ingestion, validation, transformation, data engineering, and Data MOdeling
    • Performance tune HIVE and HBASE jobs with a focus on ingestion
    • Design and develop open source platform components using Spark, Sqoop, Java, Oozie, Kafka, Python, and other components
    • Lead the technical planning & requirements gathering phases including estimate, develop, test, manage projects, architect and deliver complex projects
    • Participate and lead in design sessions, demos and prototype sessions, testing and training workshops with business users and other IT associates
    • Contribute to the thought capital through the creation of executive presentations, architecture documents and articulate them to executives through presentations
    Qualifications
    • Spring, Docker, Hibernate /Spring , JPA, Pivotal, Kafka, NoSQL,
      Hadoop Containers Docker, work, Spring boot .
    • At least 3+ years of experience on working with large projects including the most recent project in the MapR platform
    • At least 5+ years of Hands-on administration, configuration management, monitoring, performance tuning of Hadoop/Distributed platforms
    • Should have experience designing service management, orchestration, monitoring and management requirements of cloud platform.
    • Hands-on experience with Hadoop, Teradata (or other MPP RDBMS), MapReduce, Hive, Sqoop, Splunk, STORM, SPARK, Kafka and HBASE (At least 2 years)
    • Experience with end-to-end solution architecture for data capabilities including:
    • Experience with ELT/ETL development, patterns and tooling (Informatica, Talend)
    • Ability to produce high quality work products under pressure and within deadlines with specific references
    • VERY strong communication, solutioning, and client facing skills especially non-technical business users
    • At least 5+ years of working with large multi-vendor environment with multiple teams and people as a part of the project
    • At least 5+ years of working with a complex Big Data environment
    • 5+ years of experience with Team Foundation Server/JIRA/GitHub and other code management toolsets
  • Preferred Skills And Education
    Masters degree in Computer Science or related field
    Certification in Azure platform
    Perficient full-time employees receive complete and competitive benefits. We offer a collaborative work environment, competitive compensation, generous work/life opportunities and an outstanding benefits package that includes paid time off plus holidays. In addition, all colleagues are eligible for a number of rewards and recognition programs including billable bonus opportunities. Encouraging a healthy work/life balance and providing our colleagues great benefits are just part of what makes Perficient a great place to work.
    More About Perficient
    Perficient is the leading digital transformation consulting firm serving Global 2000 and enterprise customers throughout North America. With unparalleled information technology, management consulting and creative capabilities, Perficient and its Perficient Digital agency deliver vision, execution and value with outstanding digital experience, business optimization and industry solutions.
    Our work enables clients to improve productivity and competitiveness; grow and strengthen relationships with customers, suppliers and partners; and reduce costs. Perficient's professionals serve clients from a network of offices across North America and offshore locations in India and China. Traded on the Nasdaq Global Select Market, Perficient is a member of the Russell 2000 index and the S&P SmallCap 600 index.
    Perficient is an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national, origin, disability status, protected veteran status, or any other characteristic protected by law.
    Disclaimer: The above statements are not intended to be a complete statement of job content, rather to act as a guide to the essential functions performed by the employee assigned to this classification. Management retains the discretion to add or change the duties of the position at any time.
    Select work authorization questions to ask when applicants apply
    • Are you legally authorized to work in the United States?
    • Will you now, or in the future, require sponsorship for employment visa status (e.g. H-1B visa status)?