OnlyDataJobs.com

State Farm
  • Atlanta, GA

WHAT ARE THE DUTIES AND RESPONSIBILITIES OF THIS POSITION?

    Perfo
    • rms improved visual representation of data to allow clearer communication, viewer engagement and faster/better decision-making Inves
    • tigates, recommends, and initiates acquisition of new data resources from internal and external sources Works
    • with IT teams to support data collection, integration, and retention requirements based on business need Ident
    • ifies critical and emerging technologies that will support and extend quantitative analytic capabilities Manag
    • es work efforts which require the use of sophisticated project planning techniques Appli
    • es a wide application of complex principles, theories and concepts in a specific field to provide solutions to a wide range of difficult problems Devel
    • ops and maintains an effective network of both scientific and business contacts/knowledge obtaining relevant information and intelligence around the market and emergent opportunities Contr
    • ibutes data to State Farm's internal and external publications, write articles for leading journals and participate in academic and industry conferences
    • Collaborates with business subject matter experts to select relevant sources of information
    • Develop breadth of knowledge in programming (R, Python), Descriptive, Inferential, and Experimental Design statistics, advanced mathematics, and database functionality (SQL, Hadoop)
    • Develop expertise with multiple machine learning algorithms and data science techniques, such as exploratory data analysis, generative and discriminative predictive modeling, graph theory, recommender systems, text analytics, computer vision, deep learning, optimization and validation
    • Develop expertise with State Farm datasets, data repositories, and data movement processes
    • Assists on projects/requests and may lead specific tasks within the project scope
    • Prepares and manipulates data for use in development of statistical models
    • Develops fundamental understanding of insurance and financial services operations and uses this knowledge in decision making


Additional Details:

For over 95 years, data has been key to State Farm.  As a member of our data science team with the Enterprise Data & Analytics department under our Chief Data & Analytics Officer, you will work across the organization to solve business problems and help achieve business strategies.  You will employ sophisticated, statistical approaches and state of the art technology.  You will build and refine our tools/techniques and engage w/internal stakeholders across the organization to improve our products & services.


Implementing solutions is critical for success. You will do problem identification, solution proposal & presentation to a wide variety of management & technical audiences. This challenging career requires you to work on multiple concurrent projects in a community setting, developing yourself and others, and advancing data science both at State Farm and externally.


Skills & Professional Experience

·        Develop hypotheses, design experiments, and test feasibility of proposed actions to determine probable outcomes using a variety of tools & technologies

·        Masters, other advanced degrees, or five years experience in an analytical field such as data science quantitative marketing, statistics, operations research, management science, industrial engineering, economics, etc. or equivalent practical experience preferred.

·        Experience with SQL, Python, R, Java, SAS or MapReduce, SPARK

·        Experience with unstructured data sets: text analytics, image recognition etc.

·        Experience working w/numerous large data sets/data warehouses & ability to pull from such data sets using relevant programs & coding including files, RDBMS & Hadoop based storage systems

·        Knowledge in machine learning methods including at least one of the following: Time series analysis, Hierarchical Bayes; or learning techniques such as Decision Trees, Boosting, Random Forests.

·        Excellent communication skills and the ability to manage multiple diverse stakeholders across businesses & leadership levels.

·        Exercise sound judgment to diagnose & resolve problems within area of expertise

·        Familiarity with CI/CD development methods, Git and Docker a plus


Multiple location opportunity. Locations offered are: Atlanta, GA, Bloomington, IL, Dallas, TX and Phoenix, AZ


Remote work option is not available.


There is no sponsorship for an employment visa for the position at this time.


Competencies desired:
Critical Thinking
Leadership
Initiative
Resourcefulness
Relationship Building
State Farm
  • Dallas, TX

WHAT ARE THE DUTIES AND RESPONSIBILITIES OF THIS POSITION?

    Perfo
    • rms improved visual representation of data to allow clearer communication, viewer engagement and faster/better decision-making Inves
    • tigates, recommends, and initiates acquisition of new data resources from internal and external sources Works
    • with IT teams to support data collection, integration, and retention requirements based on business need Ident
    • ifies critical and emerging technologies that will support and extend quantitative analytic capabilities Manag
    • es work efforts which require the use of sophisticated project planning techniques Appli
    • es a wide application of complex principles, theories and concepts in a specific field to provide solutions to a wide range of difficult problems Devel
    • ops and maintains an effective network of both scientific and business contacts/knowledge obtaining relevant information and intelligence around the market and emergent opportunities Contr
    • ibutes data to State Farm's internal and external publications, write articles for leading journals and participate in academic and industry conferences
    • Collaborates with business subject matter experts to select relevant sources of information
    • Develop breadth of knowledge in programming (R, Python), Descriptive, Inferential, and Experimental Design statistics, advanced mathematics, and database functionality (SQL, Hadoop)
    • Develop expertise with multiple machine learning algorithms and data science techniques, such as exploratory data analysis, generative and discriminative predictive modeling, graph theory, recommender systems, text analytics, computer vision, deep learning, optimization and validation
    • Develop expertise with State Farm datasets, data repositories, and data movement processes
    • Assists on projects/requests and may lead specific tasks within the project scope
    • Prepares and manipulates data for use in development of statistical models
    • Develops fundamental understanding of insurance and financial services operations and uses this knowledge in decision making


Additional Details:

WHAT ARE THE DUTIES AND RESPONSIBILITIES OF THIS POSITION?

    Perfo
    • rms improved visual representation of data to allow clearer communication, viewer engagement and faster/better decision-making Inves
    • tigates, recommends, and initiates acquisition of new data resources from internal and external sources Works
    • with IT teams to support data collection, integration, and retention requirements based on business need Ident
    • ifies critical and emerging technologies that will support and extend quantitative analytic capabilities Manag
    • es work efforts which require the use of sophisticated project planning techniques Appli
    • es a wide application of complex principles, theories and concepts in a specific field to provide solutions to a wide range of difficult problems Devel
    • ops and maintains an effective network of both scientific and business contacts/knowledge obtaining relevant information and intelligence around the market and emergent opportunities Contr
    • ibutes data to State Farm's internal and external publications, write articles for leading journals and participate in academic and industry conferences
    • Collaborates with business subject matter experts to select relevant sources of information
    • Develop breadth of knowledge in programming (R, Python), Descriptive, Inferential, and Experimental Design statistics, advanced mathematics, and database functionality (SQL, Hadoop)
    • Develop expertise with multiple machine learning algorithms and data science techniques, such as exploratory data analysis, generative and discriminative predictive modeling, graph theory, recommender systems, text analytics, computer vision, deep learning, optimization and validation
    • Develop expertise with State Farm datasets, data repositories, and data movement processes
    • Assists on projects/requests and may lead specific tasks within the project scope
    • Prepares and manipulates data for use in development of statistical models
    • Develops fundamental understanding of insurance and financial services operations and uses this knowledge in decision making


Additional Details:

For over 95 years, data has been key to State Farm.  As a member of our data science team with the Enterprise Data & Analytics department under our Chief Data & Analytics Officer, you will work across the organization to solve business problems and help achieve business strategies.  You will employ sophisticated, statistical approaches and state of the art technology.  You will build and refine our tools/techniques and engage w/internal stakeholders across the organization to improve our products & services.


Implementing solutions is critical for success. You will do problem identification, solution proposal & presentation to a wide variety of management & technical audiences. This challenging career requires you to work on multiple concurrent projects in a community setting, developing yourself and others, and advancing data science both at State Farm and externally.


Skills & Professional Experience

·        Develop hypotheses, design experiments, and test feasibility of proposed actions to determine probable outcomes using a variety of tools & technologies

·        Masters, other advanced degrees, or five years experience in an analytical field such as data science quantitative marketing, statistics, operations research, management science, industrial engineering, economics, etc. or equivalent practical experience preferred.

·        Experience with SQL, Python, R, Java, SAS or MapReduce, SPARK

·        Experience with unstructured data sets: text analytics, image recognition etc.

·        Experience working w/numerous large data sets/data warehouses & ability to pull from such data sets using relevant programs & coding including files, RDBMS & Hadoop based storage systems

·        Knowledge in machine learning methods including at least one of the following: Time series analysis, Hierarchical Bayes; or learning techniques such as Decision Trees, Boosting, Random Forests.

·        Excellent communication skills and the ability to manage multiple diverse stakeholders across businesses & leadership levels.

·        Exercise sound judgment to diagnose & resolve problems within area of expertise

·        Familiarity with CI/CD development methods, Git and Docker a plus


Multiple location opportunity. Locations offered are: Atlanta, GA, Bloomington, IL, Dallas, TX and Phoenix, AZ


Remote work option is not available.


There is no sponsorship for an employment visa for the position at this time.


Competencies desired:
Critical Thinking
Leadership
Initiative
Resourcefulness
Relationship Building
Gravity IT Resources
  • Miami, FL

Overview of Position:

We undertaking an ambitious digital transformation across Sales, Service, Marketing, and eCommerce. We are looking for a web data analytics wizard with prior experience in digital data preparation, discovery, and predictive analytics.

The data scientist/web analyst will work with external partners, digital business partners, enterprise analytics, and technology team to strategically plan and develop datasets, measure web analytics, and execute on predictive and prescriptive use cases. The role demands the ability to (1) Learn quickly (2) Work in a fast-paced, team-driven environment (3) Manage multiple efforts simultaneously (4) Adept at using large datasets and using models to test effectiveness of different courses of action (5) Promote data driven decision making throughout the organization (6) Define and measure success of capabilities we provide the organization.


Primary Duties and Responsibilities

    Analy
    • ze data captured through Google Analytics and develop meaningful actionable insights on digital behavior. Put t
    • ogether a customer 360 data frame by connecting CRM Sales, Service, Marketing cloud data with Commerce Web behavior data and wrangle the data into a usable form. Use p
    • redictive modelling to increase and optimize customer experiences across online & offline channels. Evalu
    • ate customer experience and conversions to provide insights & tactical recommendations for web optimization
    • Execute on digital predictive use cases and collaborate with enterprise analytics team to ensure use of best tools and methodologies.
    • Lead support for enterprise voice of customer feedback analytics.
    • Enhance and maintain digital data library and definitions.

Minimum Qualifications

  • Bachelors degree in Statistics, Computer Science, Marketing, Engineering or equivalent
  • 3 years or more of working experience in building predictive models.
  • Experience in Google Analytics or similar web behavior tracking tools is required.
  • Experience in R is a must with working knowledge of connecting to multiple data sources such as amazon redshift, salesforce, google analytics, etc.
  • Working knowledge in machine learning algorithms such as Random Forest, K-means, Apriori, Support Vector machine, etc.
  • Experience in A/B testing or multivariate testing.
  • Experience in media tracking tags and pixels, UTM, and custom tracking methods.
  • Microsoft Office Excel & PPT (advanced).

Preferred Qualifications

  • Masters degree in statistics or equivalent.
  • Google Analytics 360 experience/certification.
  • SQL workbench, Postgres.
  • Alteryx experience is a plus.
  • Tableau experience is a plus.
  • Experience in HTML, JavaScript.
  • Experience in SAP analytics cloud or SAP desktop predictive tool is a plus
Signify Health
  • Dallas, TX

Position Overview:

Signify Health is looking for a savvy Data Engineer to join our growing team of deep learning specialists. This position would be responsible for evolving and optimizing data and data pipeline architectures, as well as, optimizing data flow and collection for cross-functional teams. The Data Engineer will support software developers, database architects, data analysts, and data scientists. The ideal candidate would be self-directed, passionate about optimizing data, and comfortable supporting the Data Wrangling needs of multiple teams, systems and products.

If you enjoy providing expert level IT technical services, including the direction, evaluation, selection, configuration, implementation, and integration of new and existing technologies and tools while working closely with IT team members, data scientists, and data engineers to build our next generation of AI-driven solutions, we will give you the opportunity to grow personally and professionally in a dynamic environment. Our projects are built on cooperation and teamwork, and you will find yourself working together with other talented, passionate and dedicated team member, all working towards a shared goal.

Essential Job Responsibilities:

  • Assemble large, complex data sets that meet functional / non-functional business requirements
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing data models for greater scalability, etc.
  • Leverage Azure for extraction, transformation, and loading of data from a wide variety of data sources in support of AI/ML Initiatives
  • Design and implement high performance data pipelines for distributed systems and data analytics for deep learning teams
  • Create tool-chains for analytics and data scientist team members that assist them in building and optimizing AI workflows
  • Work with data and machine learning experts to strive for greater functionality in our data and model life cycle management capabilities
  • Communicate results and ideas to key decision makers in a concise manner
  • Comply with applicable legal requirements, standards, policies and procedures including, but not limited to the Compliance requirements and HIPAA.


Qualifications:Education/Licensing Requirements:
  • High school diploma or equivalent.
  • Bachelors degree in Computer Science, Electrical Engineer, Statistics, Informatics, Information Systems, or another quantitative field. or related field or equivalent work experience.


Experience Requirements:
  • 5+ years of experience in a Data Engineer role.
  • Experience using the following software/tools preferred:
    • Experience with big data tools: Hadoop, Spark, Kafka, etc.
    • Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.
    • Experience with AWS or Azure cloud services.
    • Experience with stream-processing systems: Storm, Spark-Streaming, etc.
    • Experience with object-oriented/object function scripting languages: Python, Java, C#, etc.
  • Strong work ethic, able to work both collaboratively, and independently without a lot of direct supervision, and solid problem-solving skills
  • Must have strong communication skills (written and verbal), and possess good one-on-one interpersonal skills.
  • Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases
  • A successful history of manipulating, processing and extracting value from large disconnected datasets.
  • Working knowledge of message queuing, stream processing, and highly scalable big data data stores.
  • 2 years of experience in data modeling, ETL development, and Data warehousing
 

Essential Skills:

  • Fluently speak, read, and write English
  • Fantastic motivator and leader of teams with a demonstrated track record of mentoring and developing staff members
  • Strong point of view on who to hire and why
  • Passion for solving complex system and data challenges and desire to thrive in a constantly innovating and changing environment
  • Excellent interpersonal skills, including teamwork and negotiation
  • Excellent leadership skills
  • Superior analytical abilities, problem solving skills, technical judgment, risk assessment abilities and negotiation skills
  • Proven ability to prioritize and multi-task
  • Advanced skills in MS Office

Essential Values:

  • In Leadership Do whats right, even if its tough
  • In Collaboration Leverage our collective genius, be a team
  • In Transparency Be real
  • In Accountability Recognize that if it is to be, its up to me
  • In Passion Show commitment in heart and mind
  • In Advocacy Earn trust and business
  • In Quality Ensure what we do, we do well
Working Conditions:
  • Fast-paced environment
  • Requires working at a desk and use of a telephone and computer
  • Normal sight and hearing ability
  • Use office equipment and machinery effectively
  • Ability to ambulate to various parts of the building
  • Ability to bend, stoop
  • Work effectively with frequent interruptions
  • May require occasional overtime to meet project deadlines
  • Lifting requirements of
DISYS
  • Minneapolis, MN
Client: Banking/Financial Services
Location: 100% Remote
Duration: 12 month contract-to-hire
Position Title: NLU/NLP Predictive Modeling Consultant


***Client requirements will not allow OPT/CPT candidates for this position, or any other visa type requiring sponsorship. 

This is a new team within the organization set up specifically to perform analyses and gain insights into the "voice of the customer" through the following activities:
Review inbound customer emails, phone calls, survey results, etc.
Review data that is unstructured "natural language" text and speech data
Maintain focus on customer complaint identification and routing
Build machine learning models to scan customer communication (emails, voice, etc)
Identify complaints from non-complaints.
Classify complaints into categories
Identify escalated/high-risk complaints, e.g. claims of bias, discrimination, bait/switch, lying, etc...
Ensure routed to appropriate EO for special

Responsible for:
Focused on inbound retail (home mortgage/equity) emails
Email cleansing: removal of extraneous information (disclaimers, signatures, headers, PII)
Modeling: training models using state of art techniques
Scoring: "productionalizing" models to be consumed by the business
Governance: model documentation and Q/A with model risk group.
Implementation of model monitoring processes

Desired Qualifications:
Real-world experience building/deploying predictive models, any industry (must)
SQL background (must)
Self-starter, able to excel in fast-paced environment w/o a ton of direction (must)
Good communication skills (must)
Experience in text/speech analytics (preferred)
Python, SAS background (preferred)
Linux (nice to have)
Spark (Scala or PySpark) (nice to have)

Sentek Global
  • San Diego, CA

Sentek Global is seeking a Software Engineer to provide support to PMW 150 in San Diego, CA!


Responsibilities
  • Design, build and maintain software, develop software infrastructure and development environments, and transition older products and capabilities to the new architectures.
  • Produce effective and powerful solutions to complex problems in areas such assoftware engineering, data analytics, automation,and cybersecurity.
  • Perform analysis of existing and emerging operational and functional requirements to support the current and future systems capabilities and requirements.
  • Provide technical expertise, guidance, architecture, development and support in many different technologies directly to government customers.
  • Perform schedule planning and program management tasks as required.
  • Perform Risk Analysis for implementation of program requirements.
  • Assist in the development of requirements documents.
  • Other duties as required.


Qualifications
  • A current active secret clearance is required to be considered for this role.
  • A Bachelors Degree in data science, data analytics, computer science, or a related technical discipline is required.
  • Three to five (3-5) years providing software engineering support to a DoD program office.
  • Experience working with data rich problems through research or programs.
  • Experience with computer programming or user experience/user interface.
  • Demonstrated knowledge completing projects with large or incomplete data and ability to recommend solutions.
  • Experience with Machine Learning algorithms including convolutional neural networks (CNN), regression, classification, clustering, etc.
  • Experience using deep learning frameworks (preferably TensorFlow).
  • Experience designing and developing professional software using Linux, Python, C++, JAVA, etc.
    • Experience applying Deep/Machine Learning technology to solve real-world problems:
    • Selecting features, building and optimizing classifiers using machine learning techniques.
    • Data mining using state-of-the-art methods.
    • Extending companys data with third party sources of information when needed.
    • Enhancing data collection procedures to include information that is relevant for building analytic systems.
  • Experience processing, cleansing, and verifying the integrity of data used for analysis.
  • Experience performing ad-hoc analyses and presenting results in a clear manner.
  • Experience creating automated anomaly detection systems and constant tracking of its performance.
  • Must be able to travel one to three (1-3) times per year.
Mix.com
  • Phoenix, AZ

Are you interested in scalability & distributed systems? Do you want to work to help shaping a discovery engine powered by cutting edge technologies and machine learning at scale? If you answered yes to the above questions, Mix's Research and Development is the team for you!


In this role, you'll be part of a small and innovative team comprised of engineers and data scientists working together to understand content by leveraging machine learning and NLP technologies. You will have the opportunity to work on core problems like detection of low quality content or spam, text semantic analysis, video and image processing, content quality assessment and monitoring. Our code operates at massive scale, ingesting, processing and indexing millions of URLs.



Responsibilities

  • Write code to build an infrastructure, which is capable of scaling based on the load
  • Collaborate with researchers and data scientists to integrate innovative Machine Learning and NLP techniques with our serving, cloud and data infrastructure
  • Automate build and deployment process, and setup monitoring and alerting systems
  • Participate in the engineering life-cycle, including writing documentation and conducting code reviews


Required Qualifications

  • Strong knowledge of algorithms, data structures, object oriented programming and distributed systems
  • Fluency in OO programming language, such as  Scala (preferred), Java, C, C++
  • 3+ years demonstrated expertise in stream processing platforms like Apache Flink, Apache Storm and Apache Kafka
  • 2+ years experience with a cloud platform like Amazon Web Services (AWS) or Microsoft Azure
  • 2+ years experience with monitoring frameworks, and analyzing production platforms, UNIX servers and mission critical systems with alerting and self-healing systems
  • Creative thinker and self-starter
  • Strong communication skills


Desired Qualifications

  • Experience with Hadoop, Hive, Spark or other MapReduce solutions
  • Knowledge of statistics or machine learning
KELZAL (QELZAL CORPORATION)
  • San Diego, CA

Challenge:

As Kelzals Machine Learning Engineer, youwill be part of an innovative team that designs and develops algorithms and software for the next generation of AI-enabled visual systems. You will develop power-efficient machine learning and adaptive signal processing algorithms to solve real-world imaging and video classification problems.


Responsibilities:

  • Develop algorithms for the fast, low-complexity and accurate detection and tracking of objects in real-world environments
  • Develop algorithms for event-based spatio-temporal signal processing
  • Contribute to our machine learning tool sets for curating data and training models
  • Inform sensor decisions for optimal approaches to classification for product requirements
  • Follow and drive research on state-of-the-art approaches in the areas described above, as applied to the problems we're solving


Requirements:

·      Experience in event-based signal processing

·      Experience in continuous-time signal processing techniques

·      Experience in some deep neural network packages (e.g. TensorFlow, NVIDIA Digits,             Caffe/Caffe2)

·      Experience with OpenCV

·      Experience with traditional computer vision approaches to image processing

·      Experience with developing machine-learning algorithms for multi-modal object detection,         scene understanding, semantic classification, face verification, human pose estimation, activity recognition, or anomaly detection

·      Strong experience with classification and regression algorithms

·      Strong coding skills with Python and/or C/C++ in Linux environment

·      Track record of research excellence or/and experience converting publications to actual implementations

·      Experience with commercial development processes such as continuous integration, deployment and release management tools a plus.

·      Experience launching products containing machine learning algorithms a plus

·      Experience with fixed point implementation a plus

·      3+ years hands-on experience working in industry

·      MS or PhD Degree in Computer Science, Electrical Engineering or a related field

.      Current US work authorization

Biswas Information Technology Solutions
  • Herndon, VA

We are seeking a junior-mid level Data Science Engineer to analyze large amounts of raw data from different sources to extract valuable business insights to aid in better business decision-making. Analytical mind, problem-solving skills and passion for machine-learning and research are critical for this role. You will be part of a highly passionate development team that is in the process of refining our Data Science toolkit that includes a wide set of predictive, recommendation, and inference modeling for our AI product — ranging from time-series forecasting, sentiment analysis, custom object-detection, to named-entity recognition, text summarization, and geometric deep learning.



Responsibilities



  • Identify valuable data sources and automate collection processes

  • Preprocessing of structured and unstructured data

  • Discover trends and patterns in large amounts of data

  • Build predictive models and machine-learning algorithms

  • Present information using data visualization techniques

  • Propose solutions and strategies to business challenges

  • Collaborate with engineering and product development teams



Requirements



  • Strong fundamentals in training, evaluating, and benchmarking machine learning models

  • Strong in Python Numpy, Pandas, keras (tensorflow or pytorch is a plus)

  • Familiar with Feature selection, Feature Extraction (especially for deep learning is a plus++)

  • Familiarity of common Hyper-optimization Techniques with different AI models.

  • Experience handling large data sets

  • Familiarity with BI tools (e.g. Tableau) and data frameworks (e.g. Hadoop)

  • Strong math skills (e.g. statistics, algebra)

  • Problem-solving aptitude

  • Excellent communication and presentation skills

  • 3 to 5 years of experience in the above is preferred

UST Global
  • San Diego, CA

KEY SKILLSETS

- 7+ years experience with Python

- 4+ years experience with Java


General Responsibilities
- Selecting features, building and optimizing classifiers using machine learning techniques
- Data mining using state of the art methods
- Extending business data with third party sources of information when needed
- Enhancing data collection procedures to include information that is relevant for building analytic systems
- Processing, cleansing, and verifying the integrity of data used for analysis
- Doing ad hoc analysis and presenting results in a clear manner
- Creating automated anomaly detection systems and constant tracking of its performance
Skills and Qualifications
- Min 8 yrs of experience
- Hands on experience in Python
- Excellent understanding of machine learning techniques and algorithms.
- Experience with common data science toolkits, such as R, Weka, NumPy, MatLab, etc Excellence in at least one of these is highly desirable
- Great communication skills
- Experience with data visualization tools, such as GGplot, etc.
- Proficiency in using query languages such as SQL, Hive, Pig
- Experience with NoSQL databases, such as MongoDB
- Good applied statistics skills, such as distributions, statistical testing, regression,

UST Global
  • Atlanta, GA

KEY SKILLSETS

- 7+ years experience with Python

- 4+ years experience with Java


General Responsibilities
- Selecting features, building and optimizing classifiers using machine learning techniques
- Data mining using state of the art methods
- Extending business data with third party sources of information when needed
- Enhancing data collection procedures to include information that is relevant for building analytic systems
- Processing, cleansing, and verifying the integrity of data used for analysis
- Doing ad hoc analysis and presenting results in a clear manner
- Creating automated anomaly detection systems and constant tracking of its performance
Skills and Qualifications
- Min 8 yrs of experience
- Hands on experience in Python
- Excellent understanding of machine learning techniques and algorithms.
- Experience with common data science toolkits, such as R, Weka, NumPy, MatLab, etc Excellence in at least one of these is highly desirable
- Great communication skills
- Experience with data visualization tools, such as GGplot, etc.
- Proficiency in using query languages such as SQL, Hive, Pig
- Experience with NoSQL databases, such as MongoDB
- Good applied statistics skills, such as distributions, statistical testing, regression,

Visa
  • Austin, TX
Company Description
Common Purpose, Uncommon
Opportunity. Everyone at Visa works with one goal in mind making sure that Visa is the best way to pay and be paid, for everyone everywhere. This is our global vision and the common purpose that unites the entire Visa team. As a global payments technology company, tech is at the heart of what we do: Our VisaNet network processes over 13,000 transactions per second for people and businesses around the world, enabling them to use digital currency instead of cash and checks. We are also global advocates for financial inclusion, working with partners around the world to help those who lack access to financial services join the global economy. Visas sponsorships, including the Olympics and FIFA World Cup, celebrate teamwork, diversity, and excellence throughout the world. If you have a passion to make a difference in the lives of people around the
world, Visa offers an uncommon opportunity to build a strong, thriving career. Visa is fueled by our team of talented employees who continuously raise the bar on delivering the convenience and security of digital currency to people all over the world. Join our team and find out how Visa is everywhere you want to
be.
Job Description
The ideal candidate will be responsible for the following to:
  • Perform Hadoop Administration on Production Hadoop clusters
  • Perform Tuning and Increase Operational efficiency on a continuous basis
  • Monitor health of the platforms and Generate Performance Reports and Monitor and provide continuous improvements
  • Working closely with development, engineering and operation teams, jointly work on key deliverables ensuring production scalability and stability
  • Develop and enhance platform best practices
  • Ensure the Hadoop platform can effectively meet performance & SLA requirements
  • Responsible for support of Hadoop Production environment which includes Hive, YARN, Spark, Impala, Kafka, SOLR, Oozie, Sentry, Encryption, Hbase, etc.
  • Perform optimization capacity planning of a large multi-tenant cluster
Qualifications
  • Minimum 3 years of work experience in maintaining, optimization, issue resolution of Hadoop clusters, supporting Business users and Batch
  • Experience in Configuring and setting up Hadoop clusters and provide support for - aggregation, lookup & fact table creation criteria
  • Map Reduce tuning, data node, NN recovery etc.
  • Experience in Linux / Unix OS Services, Administration, Shell, awk scripting
  • Experience in building and scalable Hadoop applications
  • Experience in Core Java, Hadoop (Map Reduce, Hive, Pig, HDFS, H-catalog, Zookeeper and OOzie)
  • Hands-on Experience in SQL (Oracle ) and No SQL Databases (HBASE/Cassandra/Mongo DB)
  • Excellent oral and written communication and presentation skills, analytical and problem solving skills
  • Self-driven, Ability to work independently and as part of a team with proven track record developing and launching products at scale
  • Minimum of four year technical degree required
  • Experience on Cloudera distribution preferred
  • Hands-on Experience as a Linux Sys Admin is a plus
  • Knowledge on Spark and Kafka is a plus.
Additional Information
All your information will be kept confidential according to EEO guidelines.
Job Number: REF15232V
Pythian
  • Dallas, TX

Google Cloud Solutions Architect (Pre Sales)

United States | Canada | Remote | Work from Home

Why You?

Are you a US or Canada based Cloud Solutions Architect who likes to operate with a high degree of autonomy and have diverse responsibilities that require strong leadership, deep technology skills and a dedication to customer service? Do you have Big data and Data centric skills? Do you want to take part in the strategic planning of organizations data estate with a focus of fulfilling business requirements around cost, scalability and flexibility of the platform? Can you draft technology roadmaps and document best practice gaps with precise steps of how to get there? Can you implement the details of the backlogs you have helped build? Do you demonstrate consistent best practices and deliver strong customer satisfaction? Do you enjoy pre sales? Can you demonstrate adoption of new technologies and frameworks through the development of proofs of concepts?

If you have a passion for solving complex problems and for pre sales then this could be the job for you!

What Will You Be Doing?  

  • Collaborating with and supporting Pythian sales teams in the pre-sales & account management process from the technical perspective, remotely and on-site (approx 75%).
  • Defining solutions for current and future customers that efficiently address their needs. Leading through example and influence, as a master of applying technology solutions to solve business problems.
  • Developing Proof of Concepts (PoC) in order to demonstrate feasibility and value to Pythians customers (approx 25%).
  • Defining solutions for current and future customers that efficiently address their needs.
  • Identifying then executing solutions with a commitment to excellent customer service
  • Collaborating with others in refining solutions presented to customers
  • Conducting technical audits of existing architectures (Infrastructure, Performance, Security, Scalability and more) document best practices and recommendations
  • Providing component or site-wide performance optimizations and capacity planning
  • Recommending best practices & improvements to current operational processes
  • Communicating status and planning activities to customers and team members
  • Participate in periodic overtime (occasionally on short notice) travelling up to approx 50%).

What Do We Need From You?

While we realise you might not have everything on the list to be the successful candidate for the Solutions Architect job you will likely have at least 10 years experience in a variety of positions in IT. The position requires specialized knowledge and experience in performing the following:

  • Undergraduate degree in computer science, computer engineering, information technology or related field or relevant experience.
  • Systems design experience
  • Understanding and experience with Cloud architectures specifically: Google Cloud Platform (GCP) or Microsoft Azure
  • In-depth knowledge of popular database and data warehouse technologies from Microsoft, Amazon and/or Google (Big Data & Conventional RDBMS), Microsoft Azure SQL Data Warehouse, Teradata, Redshift,  BigQuery, Snowflake etc.
  • Be fluent in a few languages, preferably Java and Python, and having familiarity with Scala and Go would be a plus.
  • Proficient in SQL. (Experience with Hive and Impala would be great)
  • Proven ability to work with software engineering teams and understand complex development systems, environments and patterns.
  • Experience presenting to high level executives (VPs, C Suite)
  • This is a North American based opportunity and it is preferred that the candidate live on the West Coast, ideally in San Francisco or the Silicon Valley area but strong candidates may be considered from anywhere in the US or Canada.
  • Ability to travel and work across North America frequently (occasionally on short notice) up to 50% with some international travel also expected.

Nice-to-Haves:

  • Experience Architecting Big Data platforms using Apache Hadoop, Cloudera, Hortonworks and MapR distributions.
  • Knowledge of real-time Hadoop query engines like Dremel, Cloudera Impala, Facebook Presto or Berkley Spark/Shark.
  • Experience with BI platforms, reporting tools, data visualization products, ETL engines.
  • Experience with any MPP (Oracle Exadata/DW, Teradata, Netezza, etc)
  • Understanding of continuous delivery and deployment patterns and tools (Jenkins, Artifactory, Maven, etc)
  • Prior experience working as/with Machine Learning Engineers, Data Engineers, or Data Scientists.
  • A certification such as Google Cloud Professional Cloud Architect, Google Professional Data Engineer or related AWS Certified Solutions Architect / Big Data or Microsoft Azure Architect
  • Experience or strong interest in people management, in a player-coach style of leadership longer term would be great.

What Do You Get in Return?

  • Competitive total rewards package
  • Flexible work environment: Why commute? Work remotely from your home, theres no daily travel requirement to the office!
  • Outstanding people: Collaborate with the industrys top minds.
  • Substantial training allowance: Hone your skills or learn new ones; participate in professional development days, attend conferences, become certified, whatever you like!
  • Amazing time off: Start with a minimum 3 weeks vacation, 7 sick days, and 2 professional development days!
  • Office Allowance: A device of your choice and personalise your work environment!  
  • Fun, fun, fun: Blog during work hours; take a day off and volunteer for your favorite charity.
SoftClouds LLC
  • San Diego, CA

Job Overview: SoftClouds is looking for a Data Engineer to join our analytics platform team in designing and developing the next generation data and analytics solutions. The candidate should have deep technical skills as well as the ability to understand data and analytics, and an openness to working with disparate platforms, data sources and data formats.


Roles and Responsibilities:
  • Experience with MySQL, MS SQL Server, or Hadoop, or MongoDB.
  • Writing SQL Queries, tables joins.
  • AWS, python, or bash shell scripting
  • Have some experience pulling data from Hadoop.
  • Analyze data, system and data flows and develop effective ways to store and present data in BI applications
  • ETL experience a plus.
  • Work with data from disparate environments including Hadoop, MongoDB Talend, and other SQL and NoSQL data stores
  • Help develop the next generation analytics platform
  • Proactively ensure data integrity and focus on continuous performance improvements of existing processes.


Required skills and experience:
  • 5  or more years of experience in software development
  • 3 year of experience in writing Data applications using Spark
  • Experience in Java and Python
  • Familiarity  with Agile development methodology `
  • Experience with Scala is a plus
  • Experience with NoSQL databases, e.g., Cassandra is a plus
  • Expertise in Apache Spark & Hadoop.
  • Expertise in machine learning algorithms


Education / Experience:

  • Bachelor's Degree in Engineering or Computer Science or related field required.
  • U.S. Citizens/GC/GC EAD are encouraged to apply. We are unable to sponsor at this time. NO C2C or third-party agencies.



Booz Allen Hamilton - Tagged
  • San Diego, CA
Job Description
Job Number: R0042382
Data Scientist, Mid
The Challenge
Are you excited at the prospect of unlocking the secrets held by a data set? Are you fascinated by the possibilities presented by machine learning, artificial intelligence advances, and IoT? In an increasingly connected world, massive amounts of structured and unstructured data open up new opportunities. As a data scientist, you can turn these complex data sets into useful information to solve global challenges. Across private and public sectors-from fraud detection, to cancer research, to national intelligence-you know the answers are in the data.
We have an opportunity for you to use your analytical skills to improve the DoD and federal agencies. Youll work closely with your customer to understand their questions and needs, then dig into their data-rich environment to find the pieces of their information puzzle. Youll develop algorithms, write scripts, build predictive analytics, use automation, and apply machine learning to turn disparate data points into objective answers to help our nations services and leaders make data-driven decisions. Youll provide your customer with a deep understanding of their data, what it all means, and how they can use it. Join us as we use data science for good in the DoD and federal agencies.
Empower change with us.
Build Your Career
At Booz Allen, we know the power of data science and machine intelligence and were dedicated to helping you grow as a data scientist. When you join Booz Allen, you can expect:
  • access to online and onsite training in data analysis and presentation methodologies, and tools like Hortonworks, Docker, Tableau, Splunk, and other open source and emerging tools
  • a chance to change the world with the Data Science Bowlthe worlds premier data science for social good competition
  • participation in partnerships with data science leaders, like our partnership with NVIDIA to deliver Deep Learning Institute (DLI) training to the federal government
Youll have access to a wealth of training resources through our Analytics University, an online learning portal specifically geared towards data science and analytics skills, where you can access more than 5000 functional and technical, certifications, and books. Build your technical skills through hands-on training on the latest tools and state-of-the-art tech from our in-house experts. Pursuing certifications? Take advantage of our tuition assistance, on-site bootcamps, certification training, academic programs, vendor relationships, and a network of professionals who can give you helpful tips. Well help you develop the career you want as you chart your own course for success.
You Have
  • Experience with one or more statistical analytical programming languages, including Python or R
  • Experience with source control and dependency management software, including Git or Maven
  • Experience with using relational databases, including MySQL
  • Experience with identifying analytic insight in data, developing visualizations, and presenting findings to stakeholders
  • Knowledge of object-oriented programming, including Java and C++
  • Knowledge of various machine learning algorithms and their designs, capabilities, and limitations
  • Knowledge of statistical analysis techniques
  • Ability to build complex extraction, transformation, and loading (ETL) pipelines to clean and fuse data together
  • Ability to obtain a security clearance
  • BA or BS degree
Nice If You Have
  • Experience with designing and implementing custom machine learning algorithms
  • Experience with graph algorithms and semantic Web
  • Experience with designing and setting up relational databases
  • Experience with Big Data computing environments, including Hadoop
  • Experience with Navy mission systems
  • MA degree in Mathematics, CS, or a related quantitative field
Clearance
Applicants selected will be subject to a security investigation and may need to meet eligibility requirements for access to classified information.
Were an EOE that empowers our peopleno matter their race, color, religion, sex, gender identity, sexual orientation, national origin, disability, or veteran statusto fearlessly drive change.
, CJ1, GD13, MPPC, SIG2017
Dematic USA
  • Austin, TX
Dematic is a leading supplier of integrated automated technology, software and services to optimize the supply chain. Dematic employs over 6,000 skilled logistics professionals to serve its customers globally, with engineering centers and manufacturing facilities located around the world. Dematic is one brand under the KION Group of companies and has implemented more than 6,000 integrated systems for a customer base that includes small, medium and large companies doing business in a variety of market sectors.
Headquartered in Atlanta, Georgia, Dematic is a member of KION Group, a global leader in industrial trucks, related services and supply chain solutions. Across more than 100 countries worldwide, the KION Group designs, builds and supports logistics solutions that optimize material and information flow within factories, warehouses and distribution centers. The company is the largest manufacturer of industrial trucks in Europe, the second-largest producer of forklifts globally and a leading provider of warehouse automation.
The Role
Dematic is looking for a senior-level Operations Research Engineer in our Austin, TX office (with other locations negotiable). In Operations Research team, we design, optimize and implement intelligent and optimization ideas to the warehouse, and may stay through the complete project process from sales to development. The senior-level position should be able to approach problems by the employee himself/herself and lead a small or large team to solve problems, and be the subject expert in this field.
Key Responsibilities Include, But Not Limited To
    Sale
    • s Cycle:
    • Provide technical support to sales
    • Understand, analyze and document existing operations/requirements
    • Data parsing, processing and profiling
    • Write project proposals for distribution center operations detailing recommended solutions
    • Justify financial investment in proposed solutions
    • Present analysis findings to internal team and customer
  • Analysis/Design Cycle:
    • Process design for supply chain applications in warehouse environments
    • Develop/Design simulation models in Java/Python/Automod programming language
    • Develop and test enhanced/optimized solutions for existing processes
    • Develop and test optimization algorithms for proposed solutions
  • Product/Project Development Cycle:
    • Discuss project and algorithm design requirements with internal team and customer
    • Implement and test algorithms to production software
    • Provide technical support to developers during and after implementation cycle
  • R&D Cycle:
    • Understand the companies current product/process offering
    • Track the current process or product limitations and propose improvement plans
    • Document the current process and the proposed process
    • Communicate with stakeholders for R&D goals
    • Lead/Perform work on R&D effort, which include but not limited to:
    • Create data input
    • Model current process/product
    • Design future process/product
    • Model future process/product
  • Present findings to big group audience
  • Leadership :
    • Being technical/business lead for mid-big teams (2-7 people) that works on a project
    • Create project plans and timeline
    • Be reponsible for teams overall progress, track team memberss progress and provide technical/business suggestions
    • Make progress report to leadership teams
    • Train junior OR engineers
  • Travel to customer sites and other offices nationwide (possibly international) depending on phase of project (Travel
What We Are Looking For
    5+ y
    • ears of experience in supply chain, industrial engineering, or optimization related field Mast
    • ers degree in Operations Research, Industrial Engineering, Supply Chain, or equivalent field; OR, Bach
    • elors degree in Computer Science, Industrial Engineering, Math, Supply Chain, Logistics, Transportation, other Engineering majors or equivalent field
    • Preferred coursework/experience to include:
    • Simulation
    • Programming
    • Operations Research (Linear/Non-linear/Integer/Dynamic Programming)
    • Stochastic Processes
    • Supply Chain/Logistics/Warehousing/Transportation
  • Background/skills at:
    • Data science and machine learning/statistical approach
    • Data parsing and manipulation
    • Simulation modeling
    • Mathematical optimization (algorithm development) of processes and systems
    • Inferential statistical analysis skills
    • Linear and non-linear regression modeling
    • ROI (Return On Investment) analysis skills
    • Management/leadership skills
    • Risk analysis skills
  • Object-oriented programming skills Java (preferred), Python (preferred), or VC++
  • Data analysis skills Python (preferred), or R
  • Writing skills to create business documents (proposals, system specifications, and presentations) using both MS Word and PowerPoint
Dematic USA
  • Austin, TX
Dematic is a leading supplier of integrated automated technology, software and services to optimize the supply chain. Dematic employs over 6,000 skilled logistics professionals to serve its customers globally, with engineering centers and manufacturing facilities located around the world. Dematic is one brand under the KION Group of companies and has implemented more than 6,000 integrated systems for a customer base that includes small, medium and large companies doing business in a variety of market sectors.
Headquartered in Atlanta, Georgia, Dematic is a member of KION Group, a global leader in industrial trucks, related services and supply chain solutions. Across more than 100 countries worldwide, the KION Group designs, builds and supports logistics solutions that optimize material and information flow within factories, warehouses and distribution centers. The company is the largest manufacturer of industrial trucks in Europe, the second-largest producer of forklifts globally and a leading provider of warehouse automation.
The Role
Dematic is looking for an entry level Operations Research Engineer in our Austin, TX office (with other locations negotiable). In the Operations Research team, we design, optimize and implement intelligent and optimization ideas to the warehouse, and may stay through the complete project process from sales to development.
Key Responsibilities Include, But Not Limited To
    Sale
    • s Cycle:
    • Provide technical support to sales
    • Understand, analyze and document existing operations/requirements
    • Data parsing, processing and profiling
    • Write project proposals for distribution center operations detailing recommended solutions
    • Justify financial investment in proposed solutions
    • Present analysis findings to internal team and customer
  • Analysis/Design Cycle:
    • Process design for supply chain applications in warehouse environments
    • Develop/Design simulation models in Java/Python/Automod programming language
    • Develop and test enhanced/optimized solutions for existing processes
    • Develop and test optimization algorithms for proposed solutions
  • Product/Project Development Cycle:
    • Discuss project and algorithm design requirements with internal team and customer
    • Implement and test algorithms to production software
    • Provide technical support to developers during and after implementation cycle
  • R&D Cycle:
    • Understand the companies current product/process offering
    • Track the current process or product limitations and propose improvement plans
    • Document the current process and the proposed process
    • Communicate with stakeholders for R&D goals
    • Lead/Perform work on R&D effort, which include but not limited to:
    • Create data input
    • Model current process/product
    • Design future process/product
    • Model future process/product
  • Present findings to big group audience
    • Travel to customer sites nationwide (possibly international) depending on phase of project (Travel
What We Are Looking For
    Mast
    • ers degree (or above) in Operations Research, Industrial Engineering, Supply Chain, or equivalent field Bach
    • elors degree in Computer Science, Industrial Engineering, Math, Supply Chain, Logistics, Transportation, other Engineering majors or equivalent field Pref
    • erred coursework to include:
    • Simulation
    • Programming
    • Operations Research (Linear/Non-linear/Integer/Dynamic Programming)
    • Stochastic Processes
    • Supply Chain/Logistics/Warehousing/Transportation
  • Background/skills at:
    • Data science and machine learning/statistical approach
    • Data parsing and manipulation
    • Simulation modeling
    • Mathematical optimization (algorithm development) of processes and systems
    • Inferential statistical analysis skills
    • Linear and non-linear regression modeling
    • ROI (Return On Investment) analysis skills
    • Risk analysis skills
  • Object-oriented programming skills Java (preferred), Python (preferred), or VC++
  • Data analysis skills Python (preferred), or R
  • Writing skills to create business documents (proposals, system specifications, and presentations) using both MS Word and PowerPoint
Dematic USA
  • Austin, TX
Dematic is a leading supplier of integrated automated technology, software and services to optimize the supply chain. Dematic employs over 6,000 skilled logistics professionals to serve its customers globally, with engineering centers and manufacturing facilities located around the world. Dematic is one brand under the KION Group of companies and has implemented more than 6,000 integrated systems for a customer base that includes small, medium and large companies doing business in a variety of market sectors.
Headquartered in Atlanta, Georgia, Dematic is a member of KION Group, a global leader in industrial trucks, related services and supply chain solutions. Across more than 100 countries worldwide, the KION Group designs, builds and supports logistics solutions that optimize material and information flow within factories, warehouses and distribution centers. The company is the largest manufacturer of industrial trucks in Europe, the second-largest producer of forklifts globally and a leading provider of warehouse automation.
The Role
Dematic is looking for a mid-level Operations Research Engineer in our Austin, TX office (with other locations negotiable). In the Operations Research team, we design, optimize and implement intelligent and optimization ideas to the warehouse, and may stay through the complete project process from sales to development. The mid-level position should be able to approach problems by the employee himself/herself and lead a small team to solve problems.
Key Responsibilities Include, But Not Limited To
    Sale
    • s Cycle:
    • Provide technical support to sales
    • Understand, analyze and document existing operations/requirements
    • Data parsing, processing and profiling
    • Write project proposals for distribution center operations detailing recommended solutions
    • Justify financial investment in proposed solutions
    • Present analysis findings to internal team and customer
  • Analysis/Design Cycle:
    • Process design for supply chain applications in warehouse environments
    • Develop/Design simulation models in Java/Python/Automod programming language
    • Develop and test enhanced/optimized solutions for existing processes
    • Develop and test optimization algorithms for proposed solutions
  • Product/Project Development Cycle:
    • Discuss project and algorithm design requirements with internal team and customer
    • Implement and test algorithms to production software
    • Provide technical support to developers during and after implementation cycle
  • R&D Cycle:
    • Understand the companies current product/process offering
    • Track the current process or product limitations and propose improvement plans
    • Document the current process and the proposed process
    • Communicate with stakeholders for R&D goals
    • Lead/Perform work on R&D effort, which include but not limited to:
    • Create data input
    • Model current process/product
    • Design future process/product
    • Model future process/product
  • Present findings to big group audience
  • Leadership :
    • Being technical/business lead for small teams (2-3 people) that works on a project
    • Create project plans and timeline
    • Be reponsible for teams overall progress, track team memberss progress and provide technical/business suggestions
    • Make progress report to leadership teams
  • Travel to customer sites and other offices nationwide (possibly international) depending on phase of project (Travel
What We Are Looking For
    2+ y
    • ears of experience in supply chain, industrial engineering, or optimization related field, with masters degree in Operations Research, Industrial Engineering, Supply Chain, or equivalent field; OR, 3+ y
    • ears of experience in supply chain, industrial engineering, or optimization related field, with bachelors degree in Computer Science, Industrial Engineering, Math, Supply Chain, Logistics, Transportation, other Engineering majors or equivalent field
    • Preferred coursework/experience to include:
    • Simulation
    • Programming
    • Operations Research (Linear/Non-linear/Integer/Dynamic Programming)
    • Stochastic Processes
    • Supply Chain/Logistics/Warehousing/Transportation
  • Background/skills at:
    • Data science and machine learning/statistical approach
    • Data parsing and manipulation
    • Simulation modeling
    • Mathematical optimization (algorithm development) of processes and systems
    • Inferential statistical analysis skills
    • Linear and non-linear regression modeling
    • ROI (Return On Investment) analysis skills
    • Management/leadership skills
    • Risk analysis skills
  • Object-oriented programming skills Java (preferred), Python (preferred), or VC++
  • Data analysis skills Python (preferred), or R
  • Writing skills to create business documents (proposals, system specifications, and presentations) using both MS Word and PowerPoint
phData, Inc.
  • Minneapolis, MN

Title: Big Data Solutions Architect (Minneapolis or US Remote)


Join the Game-Changers in Big Data  


Are you inspired by innovation, hard work and a passion for data?    


If so, this may be the ideal opportunity to leverage your background in Big Data and Software Engineering, Data Engineering or Data Analytics experience to design, develop and innovate big data solutions for a diverse set of clients.  


As a Solution Architect on our Big Data Consulting team, your responsibilities include:


    • Design, develop, and innovative Big Data solutions; partner with our internal Managed Services Architects and Data Engineers to build creative solutions to solve tough big data problems.  
    • Determine the project road map, select the best tools, assign tasks and priorities, and assume general project management oversight for performance, data integration, ecosystem integration, and security of big data solutions
    • Work across a broad range of technologies from infrastructure to applications to ensure the ideal Big Data solution is implemented and optimized
    • Integrate data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures (AWS); determine new and existing data sources
    • Design and implement streaming, data lake, and analytics big data solutions

    • Create and direct testing strategies including unit, integration, and full end-to-end tests of data pipelines

    • Select the right storage solution for a project - comparing Kudu, HBase, HDFS, and relational databases based on their strengths

    • Utilize ETL processes to build data repositories; integrate data into Hadoop data lake using Sqoop (batch ingest), Kafka (streaming), Spark, Hive or Impala (transformation)

    • Partner with our Managed Services team to design and install on prem or cloud based infrastructure including networking, virtual machines, containers, and software

    • Determine and select best tools to ensure optimized data performance; perform Data Analysis utilizing Spark, Hive, and Impala

    • Mentor and coach Developers and Data Engineers. Provide guidance with project creation, application structure, automation, code style, testing, and code reviews

Qualifications

  • 5+ years previous experience as a Software Engineer, Data Engineer or Data Analytics - combined with an expertise in Hadoop Technologies and Java programming
  • Technical Leadership experience leading/mentoring junior software/data engineers, as well as scoping activities on large scale, complex technology projects
  • Expertise in core Hadoop technologies including HDFS, Hive and YARN.  
  • Deep experience in one or more ecosystem products/languages such as HBase, Spark, Impala, Solr, Kudu, etc
  • Expert programming experience in Java, Scala, or other statically typed programming language
  • Strong working knowledge of SQL and the ability to write, debug, and optimize distributed SQL queries
  • Excellent communication skills including proven experience working with key stakeholders and customers
  • Ability to translate big picture business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access and consumption, as well as custom analytics
  • Customer relationship management including project escalations, and participating in executive steering meetings
  • Ability to learn new technologies in a quickly changing field
phData, Inc.
  • Minneapolis, MN

Title: Big Data Solutions Architect (Minneapolis or US Remote)


Join the Game-Changers in Big Data  


Are you inspired by innovation, hard work and a passion for data?    


If so, this may be the ideal opportunity to leverage your background in Big Data and Software Engineering, Data Engineering or Data Analytics experience to design, develop and innovate big data solutions for a diverse set of clients.  


As a Solution Architect on our Big Data Consulting team, your responsibilities include:


    • Design, develop, and innovative Big Data solutions; partner with our internal Managed Services Architects and Data Engineers to build creative solutions to solve tough big data problems.  
    • Determine the project road map, select the best tools, assign tasks and priorities, and assume general project management oversight for performance, data integration, ecosystem integration, and security of big data solutions
    • Work across a broad range of technologies from infrastructure to applications to ensure the ideal Big Data solution is implemented and optimized
    • Integrate data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures (AWS); determine new and existing data sources
    • Design and implement streaming, data lake, and analytics big data solutions

    • Create and direct testing strategies including unit, integration, and full end-to-end tests of data pipelines

    • Select the right storage solution for a project - comparing Kudu, HBase, HDFS, and relational databases based on their strengths

    • Utilize ETL processes to build data repositories; integrate data into Hadoop data lake using Sqoop (batch ingest), Kafka (streaming), Spark, Hive or Impala (transformation)

    • Partner with our Managed Services team to design and install on prem or cloud based infrastructure including networking, virtual machines, containers, and software

    • Determine and select best tools to ensure optimized data performance; perform Data Analysis utilizing Spark, Hive, and Impala

    • Mentor and coach Developers and Data Engineers. Provide guidance with project creation, application structure, automation, code style, testing, and code reviews

Qualifications

  • 5+ years previous experience as a Software Engineer, Data Engineer or Data Analytics - combined with an expertise in Hadoop Technologies and Java programming
  • Technical Leadership experience leading/mentoring junior software/data engineers, as well as scoping activities on large scale, complex technology projects
  • Expertise in core Hadoop technologies including HDFS, Hive and YARN.  
  • Deep experience in one or more ecosystem products/languages such as HBase, Spark, Impala, Solr, Kudu, etc
  • Expert programming experience in Java, Scala, or other statically typed programming language
  • Strong working knowledge of SQL and the ability to write, debug, and optimize distributed SQL queries
  • Excellent communication skills including proven experience working with key stakeholders and customers
  • Ability to translate big picture business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access and consumption, as well as custom analytics
  • Customer relationship management including project escalations, and participating in executive steering meetings
  • Ability to learn new technologies in a quickly changing field