OnlyDataJobs.com

Elan Partners
  • Dallas, TX

Title: ETL Developer (Data Engineer)


The data engineer is someone who develops, constructs, tests and maintains architectures, such as databases and large-scale processing systems. Clean, transform, and organize data.

The data engineers will need to recommend and sometimes implement ways to improve data reliability, efficiency, and quality. Employ a variety of languages and tools to marry systems together or try to hunt down opportunities to acquire new data from other systems and assist in further processing by other data engineers, scientists, analysts.

Ensure that the architecture that is in place supports the requirements of the data scientists and the stakeholders, and the business.


Responsibilities:

- Develop, construct, test and maintain architectures

- Ensure architecture will support the requirements of the business

- Discover opportunities for data acquisition

- Develop data set processes for data modeling, mining and production

-Employ a variety of languages and tools to marry systems together

- Recommend ways to improve data reliability, efficiency and quality

-Employ sophisticated analytics programs, machine learning and statistical methods to prepare data for use in predictive and prescriptive modeling

Required Skills: -5+ years Data Engineering/ETL Development
-Experience with Snaplogic and/or Mulesoft strongly desired
-Experience with traditional ETL tools such as SSIS, Informatica and/or DataStage
-Programming and/or Scripting experience with Python, Spark, etc.
-Exposure to Cloud (AWS and/or Azure)
-Experience with other technologies such as MDM, PowerBI, Microservices,
 Snowflake, and Big Data Tool sets highly desired

Benefits of working for our Client:
-Excellent benefits
-Bonus program
-Collaborative, friendly work environment!

JHU Applied Physics Laboratory
  • Laurel, MD
  • Salary: $100k - 140k

The Johns Hopkins Applied Physics Laboratory (APL) is a national leader in scientific research and development. APL is actively seeking a Senior Data Scientist for the Health Data Sciences & Analytics Group. The Senior Data Scientist will support the National Health Mission Area, whose aim is to revolutionize health through science and engineering. JHU/APL is located midway between Baltimore and Washington, DC.


The Health Data Science and Analytics Group provides cutting edge analytics contributions to national and global publichealth and healthcare challenges, developing solutions in Health Informatics, Population Health, Precision Medicine, Digital Health, Analytics and Software Systems. Our multidisciplinary team of engineers and scientists develop statistical and machine learning algorithms and incorporate visual analytics into software systems that process large and complex data sets. We are looking for data scientists, computer scientists, applied mathematicians, statisticians and software developers that are creative problem solvers and team players dedicated to building world class expertise to provide
solutions for health and healthcare systems around the globe.


Job Summary:
Design and develop novel computational algorithms and statistical methods and design corresponding data architectures to analyze large and complex data for a variety of challenging health and healthcare problems.
Duties:
1. Develop advanced algorithms and create software applications to perform analytics on large-scale and complex data for real-world health and healthcare applications. Promote a climate conducive to intellectual curiosity, creativity, innovation, collaboration, growth, life-long learning, productivity, and respect for others.
2. Be a leader in data science and analytics efforts. Provide input to team leads and other analysts to help define the team’s vision, design and execute analytic projects from start-to-finish, inform technical direction, and support reporting of accomplishments. Assure milestones are met on time and be responsive to sponsor needs. Build collaboration among health stakeholders, working across organizations to bring consensus to achieve objectives. Become a sought out resource by consistently producing high-quality results.
3. Document and present papers to communicate impact of research and engage with sponsor and stakeholder community.
4. Communicate often and effectively with team, sponsors and JHU/APL leadership. Participate in the data science, analytics and APL community. Take advantage of collaboration and innovation opportunities to help ensure success of APL’s mission.


Note: This job summary and listing of duties is for the purpose of describing the position and its essential functions at time of hire and may change over time.


Qualifications - External
Required Qualifications:
• M.S. in Computer Science, Information Science, Mathematics, Statistics, Data Science, or related field.
• 5-10+ years of experience.
• Demonstrated ability in selecting, developing, and applying machine learning and data mining algorithms.
• Working knowledge of modern large-scale data systems and architectures; ability to manage and manipulate large disparate data sets.
• Experience with graph analytics.
• Experience with pattern recognition, statistical analysis and machine learning; fluent, with hands-on experience with some of the following implementation languages: Python, R, Matlab, JAVA, or C++/C;
• Excellent interpersonal skills and outstanding written and oral communication skills; ability to articulate complex technical issues effectively and appropriately for a wide range of audiences.
• Strong problem solving skills strong analytical and organizational skills; ability to work independently or within a group.
• Must be eligible for Secret clearance requiring background investigation.


Desired Qualifications:
• Ph.D. in the disciplines listed above.
• Demonstrated capability to carry out original machine learning research beyond incremental application of existing techniques, as evidenced by publications in premier conferences.
• Research records that illustrate in-depth understanding of underlying theory necessary to develop novel algorithms to address unique real-world challenges.
• Extensive experience in developing and applying machine learning algorithms in health and healthcare application settings.
• Research experience with advanced machine learning research topics.
• Experience with data–driven predictive model development, unstructured text mining, natural language processing, and anomaly and novelty detection.
• A strong technical writing background.
• Experience in medicine, emergency response, or public health applications and/or exposure to clinical information systems and medical data standards.


Special Working Conditions: Some travel to local sponsor sites and support for field testing may be required.



Security: Applicant selected will be subject to a government security clearance investigation and must meet the requirements for access to classified information. Eligibility requirements include U.S. citizenship.


Equal Employment Opportunity: Johns Hopkins University/Applied Physics Laboratory (APL) is an Equal Opportunity/Affirmative Action employer that complies with Title IX of the Education Amendments Acts of 1972, as well as other applicable laws. All qualified applicants will receive consideration for employment without regard to race, color, religion, sexual orientation, gender identity, national origin, disability, or protected Veteran status.

The Wellcome Trust Sanger Institute
  • Cambridge, UK




Salary range: £36,000-£44,000 per annum depending on experience plus excellent benefits. Fixed Term Contract for 3 Years.

Open Targets has recently launched Open Targets Genetics (https://genetics.opentargets.org), a portal that aggregates large scale GWAS data with functional genomics data to identify potential drug targets at disease-associated loci.

A Statistical Geneticist role funded by Open Targets, is available at the Wellcome Sanger Institute in a new team under the leadership of Dr. Maya Ghoussaini. This is an exciting opportunity for you to participate in the enhancement of the existing Open Targets Genetics Portal through de development of new functionality and features.

You will actively engage in the integration of new eQTL datasets and tissue-specific chromatin interaction datasets.

You will have the opportunity to work across a range of analysis such as:
  • Aggregate large scase GWAS data from multiple consortia and across a wide range of disease and traits.
  • Perform association analysis on UK Biobank data with a particular focus on therapeutic areas important for Open Targets
  • Work together with other members of the Open Targets team on statistical genetics analysis for large scale sequence analysis
  • Work with existing members of the team to integrate genetic and cell-specific genomic data to identify and validate causal links between targets and diseases and improve the Genetics Portal.


We welcome candidates with a background in statistical genetics or relevant discipline with advanced level of programming skills suitable for statistical genetic analyses of complex diseases. Experience in functional genomics data analysis is highly desirable. You will have the opportunity to interact with active computational and experimental research teams using cutting edge genomic techniques.

Essential Skills

  • PhD in Statistical Genetics, Computational Biology or a closely related discipline.
  • Advanced level programming skills suitable for statistical genetic analyses, such as R, Python, MATLAB.
  • Firm grounding in statistical methods of complex disease genetics such as genome wide association studies, fine-mapping, high-throughput expression data, whole exome/genome sequencing, PheWAS, Mendelian Randomisation.
  • Previous experience in working with large-scale genetic datasets.
  • Ability to work to tight timelines.
  • Demonstrable good project management and organisational skills.
  • Fluent in written and spoken English.
  • Ability to communicate ideas and results effectively.
  • Ability to work independently and organise own workload.


Ideal Skills

  • Experience in functional genomics data analysis (RNAseq, ChIPseq, etc);
  • Experience with generating reproducible bioinformatics pipelines;
  • A strong track record in preparing publications and other written materials;
  • Interest in target validation and translational research.


Other information



Open Targets is a pioneering public-private initiative between GlaxoSmithKline (GSK), Biogen, Takeda, Celgene, Sanofi, EMBL-EBI (European Bioinformatics Institute) and the WSI (Wellcome Sanger Institute), located on the Wellcome Genome Campus in Hinxton, near Cambridge, UK.

Open Targets aims to generate evidence on the biological validity of therapeutic targets and provide an initial assessment of the likely effectiveness of pharmacological intervention on these targets, using genome-scale experiments and analysis. Open Targets aims to provide an R&D framework that applies to all aspects of human disease, to improve the success rate for discovering new medicines and share its data openly in the interests of accelerating drug discovery.

Genome Research Limited is an Equal Opportunity employer. As part of our dedication to gender equality and promoting women's careers in science, we hold an Athena SWAN Bronze Award. We will consider all qualified applicants without discrimination on grounds of disability, sexual orientation, pregnancy or maternity leave status, race or national or ethnic origin, age, religion or belief, gender identity or re-assignment, marital or civil partnership status, protected veteran status (if applicable) or any other characteristic protected by law.

Please include a covering letter and CV with your application

Closing date: 28th February, however applications will be reviewed on an ongoing basis and therefore the post may be filled before the deadline.
Talent
  • Houston, TX

We need strong technical expertise in Data Science, but beyond that this is an opportunity to help us setup a best-practice data science process, to help us determine the direction of future tooling, and to be a central part of a team that will spearhead how the company engages in Data Science.


What you'll be doing:

You will work with other Data Scientists, Data Engineers, Service Designers and Project Managers on interdisciplinary projects, using Mathematics, Statistics and Machine Learning to derive structure and knowledge from raw data.

You are a highly collaborative individual who will challenge others in your team thoughtfully while prioritising impact. You believe in iterative change, experimenting with new approaches, and learning from and teaching others.


What we are looking for: 

  • 5+ years experience working with and analysing large data sets.
  • Export knowledge of statistics.
  • Real-world experience in working with product and business teams to: identify important questions and data needs, and apply statistical methods to data to find answers.
  • Strong knowledge of Python and relevant libraries (PySpark, Pandas, etc).
  • The ability to communicate results clearly and a focus on driving impact.
  • An inquisitive nature in diving into data inconsistencies to pinpoint issues.
  • Proficiency at driving the collection of new data and refining existing data sources.
  • Excellent presentation and communication skills, with the ability to explain complex analytical concepts to people from other fields.


What will make you stand out:

  • A PhD or MS in a quantitative field (e.g., Economics, Statistics, Computer Science, Sciences, Engineering, Mathematics).
  • Prior experience with writing and debugging data pipelines using a distributed data framework (Spark, etc).
  • Best practices in software development and in productionising data science.


Bonuses to include as part of your application:

  • Links to online profiles you use such as Github, Twitter, etc.
  • A description of the most interesting data analysis youve done, its key findings and its impact.
  • A link or attachment of code youve written related to data analysis.
  • A description of your work history (whether as a resume, or LinkedIn profile).

Charter Solutions
  • Minneapolis, MN

We are looking for a Lead Data Scientist and Data Scientist that is excited to join an entrepreneurial technical team that are applying innovative analytic approaches to extremely complex problems in the Industrial Internet of Things (IIoT) space.  

Direct Permanent Hire

WhiteBear Lake ,MN

This position will be responsible for building predictive models in support of our mission-critical digital strategy initiatives, while creating the sustaining modeling and process governance to support the long-term health of team.

Required

    • 8-15 yrs experience
    • Experience working with high velocity machine data / time series data is required
    • Experience with the full life cycle of model development and deployment; hands-on experience working with application teams to deploy models in real-time or near-real-time environments
    • Bachelors or Masters in Computer Science /Masters or PhD in Computer Science, Data Science or a statistics-related discipline

Qualifications/Skills

    • Experience in predictive analytics capabilities by designing
    • Develop performance models using machine learnings for any instrumented piece of equipment, including components and subsystems (compressors, heat exchangers, pumps, fans, etc)

·    Test and evaluate the quality of algorithms using statistical methods

    • Expert knowledge of R or Python, SQL
    • Demonstrated (real world) experience with:
      • Using statistical inference, unsupervised and supervised machine learning techniques and specialized methods (e.g. reliability / survival models) in real world applications
      • Neural networks, cross validation, and hyperparameter tuning
      • Understanding of cloud infrastructure (AWS/Redshift) in the model development process
    • Experience with analytics deployment platforms

Ask for a detailed job description if you are interested. Would be glad to share more details. 

JoCo
  • Houston, TX

What is the position:


The Sr. Data Analyst will be responsible for managing master data sets, troubleshooting data issues, and developing reports.


What will you do:


  • Develop, implement, and maintain analytic systems
  • Develop and maintain protocols for handling, processing, and cleaning data
  • Evaluate complicated problems and build simple frameworks
  • Identify trends and opportunities for growth through analysis of complex data sets
  • Work with management and users to gather requirements and provide status updates
  • Work with business owners to understand their analytical needs, including identifying critical metrics and KPIs, and deliver actionable insights to relevant decision-makers
  • Evaluate organizational methods and provide source-to-target mappings and information-model specification documents for data sets
  • Evaluate internal systems for efficiency, inaccuracies, and problems
  • Create best-practice reports based on data mining, analysis, and visualization
  • Use statistical methods to analyze data and create useful reports


What are the requirements:


  • Bachelors Degree in CS, Statistics, Information Systems, or related
  • 3+ years experience with data mining
  • Experience in the Oil & Gas Industry preferred
  • Strong experience working with data discovery, analytics, and BI software tools (ie. Tableau, Qlik, PowerBI, etc.)
  • Experience with technical writing in relevant areas, including queries and reports
  • Experience with advanced analytics tools for Object-oriented scripting using languages such as (R, Python, Java, C++, etc.)
  • Experience working with SQL on Hadoop tools and technologies (ie. HIVE, Impala, Presto, Hortonworks Data Flow (HDF), Dremio, Informatica, Talend)
  • Experience with database programming languages including SQL, PL/SQL, etc.
  • Knowledge of NoSQL/Hadoop oriented databases (MongoDB, Cassandra, etc.)
  • Excellent communication skills


You would be really happy working here if:


  • Roadblocks dont intimidate you. You understand how to successfully evaluate problems and develop appropriate solutions.
  • You can be counted on in crucial times, possessing great focus while completing projects successfully and efficiently.
Roche in Germany
  • Penzberg, Germany
  • Salary: €77k - 99k

What you need to know about the department:


This position is located at the department of biostatistics. You work in a highly qualified team within R&D of Roche Diagnostics. The team consists of 7 people and in the department you find 30 colleagues, who are statisiticians, physicists, bioinformaticians and epidimiologists for example. The project you would work on is focused on the development and maintenance of an R/shiny-based platform providing statistical solutions to internal and external customers aiming at evaluating the technical and clinical performance of in-vitro diagnostic assay. Each Roche employee is a potential customer of our intranet-application, although, it is focused on R&D test-development and clinical operations. Methods and solutions will be optimized to these customers’ needs with regard to usability and quality. Also see www.Code4life.com.


What will be your daily tasks:



  • You will maintain, further develop, and optimize the R/shiny based intranet-platform

  • You will manage as Development Lead the project in terms of user requirements and specifications

  • You will specify algorithms and software solutions for frequent use cases and build first prototypes

  • You will design solutions for a wide range of algorithms and data science applications, including the design of graphical user interfaces (GUI)

  • You will develop a library of test cases to proof the correctness of a module and apply them in an automated fashion

  • You will give trainings on how to apply your software and interpret the results

  • You will analyze existing processes and model them mathematically, using simulations and optimizations in order to improve them

  • You will continuously improve your expertise by attending workshops, congresses etc.


Who you are, when you have fun to work in a high motivated team:



  • You hold a degree, ideally a doctorate or equivalent experience, in mathematics, statistics, (bio)informatics or comparable

  • You are an expert in R/shiny, Python Flask or similar framework programming

  • You have much experience with web-technologies (HTML/CSS/jQuery), XML, Linux- and web-server configuration

  • You are familiar with standard methods of software development, like version control and change-management

  • You have strong knowledge of biostatistics, applied mathematics and numerics

  • You may have already applied biometrical and statistical methods in the industry

  • Basic knowledge in biochemistry and medical terminology is an advantage

  • You have proficient language skills, both in German and English (in written and spoken form)

  • You enjoy working in (interdisciplinary) teams 


The position is subject to a time limit due to projects for a maximum of 3 years.


Get in touch



  1. Seifert +498856/607724



  • For questions, you can use our FAQ

  • If the FAQ does not answer your questions, please contact our AskHR team
    +49 621 759 74999 (external) / 3372 (internal) for technical questions.


Should you have recruiting specific questions on your application or the specific job posting, please contact the responsible recruiter K. Seifert


Application documents


For a complete application please add an updated CV, all relevant certificates and job references. Please consider that after submission you won’t be able to add any further documents.

MailChimp
  • Atlanta, GA
Mailchimp is a leading marketing platform for small business. We empower millions of customers around the world to build their brands and grow their companies with a suite of marketing automation, multichannel campaign, CRM, and analytics tools.
The Growth team at Mailchimp uses data-driven experimentation to help Mailchimp users get more value out of our products and drive toward achieving our company goals. You will be responsible for producing meaningful analytics to inform the growth process and serve as the subjecting matter expect in all matters related to A/B and MVT test analysis for the growth product team. Youre a skilled collaborator whos able to work across departments and disciplines. You can communicate the value of data-driven experimentation to other cross functional team members. The ideal candidate will have experience in similar marketing optimization, product analytics, or growth analytics role with strong technical and analytical abilities.
As a part of the Growth team, the work you do will have high visibility, as we quickly turn insights into action and drive change on Mailchimp.com and within the Mailchimp product. If this sounds like you, we would love to hear from you!
Responsibilities
    • Translate customer and business needs into actionable analytics that inform Growth strategy and generate test ideas to help meet the goals of the business; Independently act on your recommendations and deliver key insights to the team
    • Partner with qualitative research to work on better understanding the customer journey, provide quantitative insights that help inform the customer journey, and work together to complete analysis and generate test ideas
    • Be a subject matter expert within cross functional Growth Product team, including upholding best practices around A/B testing and educating others on key concepts such as sample size estimation, confidence intervals, and statistical significance
    • Take ideas put forward by the team and create a hypothesis that captures what the team is trying to learn, has clear and measurable KPIs, and can be tested in a reasonable amount of time
    • Provide regular updates and generate quantitative results to show how the Growth team is making progress towards their teams KPI/North Star Metric
    • Create and maintain automated reporting and dashboards to track key marketing and customer experience metrics across multiple properties leveraging SQL, Google BigQuery, and Google Data Studio; monitor for changes in trends, share insights and make recommendations
    • Create requirements for data tracking needs to ensure the hypothesis can be accurately measured and reported on at the end of a test

Requirements
    • Extensive experience in an analytics focused role, product analytics experience is a plus
    • Bachelors or Graduate degree (business or mathematics a plus or equivalent work experience)
    • Desire to work in a fast-paced environment
    • Expertise in A/B testing analytics and best practices, experience with Optimizely is a plus
    • Expertise in SQL, Web Analytics, Excel - R/Python is a plus
    • Proficiency in wrangling and transforming data
    • Strong communication, collaboration, and problem-solving abilities
    • Demonstrated, hands-on experience with data visualization tools
    • Expertise with web analytics tools, specifically Google Analytics (certification preferred)
    • Experience working with Google BigQuery a plus
    • Experience with statistical methods such as regression and hypothesis testing
    • Proven experience analyzing data from a variety of different sources (quantitative and qualitative), presenting the data in a clear and concise manner, and create actionable insights

Mailchimp is a founder-owned and highly profitable company headquartered in the heart of Atlanta. Our purpose is to empower the underdog, and our mission is to democratize cutting edge marketing technology for small business. We offer our employees an exceptional workplace , extremely competitive compensation, fully paid benefits (for employees and their families), and generous profit sharing . We hire humble , collaborative, and ambitious people, and give them endless opportunities to grow and succeed.
We love our hometown and support sustainable urban renewal. Our headquarters is in the historic Ponce City Market , right on the Atlanta Beltline . If you'd like to be considered for this position, please apply below. We look forward to meeting you!
Mailchimp is an equal opportunity employer, and we value diversity at our company. We don't discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.
Expedia Group - Europe
  • Genève, Switzerland

Would you like to solve real world problems using advanced machine learning for the largest online travel company worldwide?

The Expedia Group team in Geneva helps our customers and partners to find and post the best accommodation supply on our travel platform, by transforming raw data into insights that power external and internal services.

This role is for a Senior Data Scientist to join our fast-growing Data Science team.  Our mission is to help the organization understand, shape and optimize its hotel and vacation rental portfolio across the world.

We create state of the art machine learning models to make our travel platform more efficient and increase its relevance for travelers and hotel partners worldwide, for example by optimizing our lodging supply and forecasting demand.

Everything we do has measurable business impact. We build innovative algorithms and models that make intelligent, automated decisions, in batch and in real time, over most aspects of our lodging platform. We collaborate closely with the analytics, market management, product and technology teams.

In this position, you will drive key efforts of modelling lodging demand and supply, using data on millions of daily transactions across more than a million properties worldwide. You will dive into state-of-the-art machine learning, deep learning and statistical modeling approaches, experiment and apply at scale to impact the business. You will use the latest cloud and data technologies to train and deploy your models and power multiple downstream applications across the organization and support key business initiatives. Outstanding test-and-learn culture is essential to thrive in this fast-moving industry; if you can prove the positive impact of your approaches, you will quickly see them in production.

You are technically strong and possess excellent business acumen and communication skills. You can handle both planned and ad-hoc work, own important data products, prioritize workload effectively, and thrive in a dynamic environment. You are self-motivated, a fast learner and work well under pressure to meet deadlines. Communicating your findings and implications to your teammates and to business partners clearly and concisely is key.


What you'll do:



  • Understand business challenges and formalize them into appropriate machine learning frameworks

  • Gather and manipulate large volume of structured or unstructured data

  • Develop, assess, and iteratively improve algorithms based on state-of-the-art machine learning, deep learning and statistical methods

  • Collaborate extensively with business partners, program management, and engineering teams to ensure that solutions meet business needs and have functional feasibility and robustness

  • Communicate in a clear and concise manner to your peers and business stakeholders

  • Develop a sound understanding of the business, the data and the technology stack to find new business applications of data science and challenge the thinking of people around you

  • Keep abreast of latest data science development and coach the team on them


Who you are:



  • Hands-on experience with a large array of machine learning algorithms, econometrics methods and/or Bayesian statistics

  • Expertise in Python, R or Scala; good programming practice, ability to write readable, fast code

  • Knowledge in deep learning algorithms and libraries (Keras, TensorFlow, Apache MXNET, …) a plus

  • Good understanding of data technologies; Hadoop, Hive, Spark, and standard relational database structures along with query languages (SQL); experience with the cloud services (AWS, Qubole, Databricks) a plus

  • Familiarity with software engineering best practices, including version control, release management, incremental delivery, test-driven development, unit testing etc.

  • Ability to understand a business problem, identify key challenges, formalize the problem from a data and algorithm perspective, and prototype solutions

  • High intellectual curiosity and willingness to tackle complex/technical problems to generate actionable business insights

  • Strong verbal and visual communication skills


Experience and Qualification:



  • MSc or PhD (preferred) in a quantitative field like machine learning, computer science, statistics, applied mathematics/physics + some relevant industry seniority (3+ years)

  • Proven track record of applying innovative machine learning algorithms to business problems, building on state-of-the-art methods and tools

  • Proven track record of working in projects involving cross-functional teams

  • Hands-on experience with demand forecasting, time-series and count processes a plus


Relocation (Y/N): Y

Expedia is committed to creating an inclusive work environment with a diverse workforce. All qualified applicants will receive consideration for employment without regard to race, religion, gender, sexual orientation, national origin, disability or age.

Crescendo Technology Ltd.
  • Toronto, ON, Canada
About the role:  The Data Scientist will be responsible for collecting, storing, cleaning, visualising, analysing, interpreting and modelling data relevant to sports betting markets in order to recognise patterns and trends in betting markets. 


Our ideal candidate should have:


  • Expertise in R and other statistical programming languages (3+ years)

  • Minimum 2 years of experience with Frequentist and Bayesian statistical methods

  • Experience working with machine learning algorithms, probabilistic models and/or other statistical modelling approaches

  • Solid experience with modern R packages such as dplyr, ggplot2, data.table

  • Solid experience in front-end R technologies for data products such as ShinyR/FlexDashboards

  • Ability to write complex SQL queries

  • Good software engineering design principles, such as design patterns, and be able to write code to solve problems

  • Proficiency in written and spoken English


This is a key role within the team and would suit someone who is passionate about working with data/data science. We are looking for someone with strong background in statistics, modelling and algorithms (machine learning or other) and who has the ability to convey complex information through data visualisation. A thorough understanding and passion for sports and sports betting markets is ideal. Experience with cloud computing as well as python a plus.

SocialCops
  • New Delhi, India
  • Salary: ₹180k - 300k

Our Alternative Data Team builds to solve world's most critical problems. From satellite data to government reports, from structured, internal data to unstructured, external data, and from online PDFs to paper surveys, the data sources we use are broad and varied. However, what makes us different is that we don’t just sell data. Instead, we sell insights.  We integrate 200+ global data sources across different sectors - agriculture, demography, infrastructure and consumer affluence, for targeted, granular insights As a Data Science Intern at SocialCops you would get to deal with diverse data, ranging from satellite data to sales data of big companies. You will be responsible for data modelling, cleaning, structuring, and handling data sets under the mentorship of our data scientists and economists. As a Data Science Intern, you will play a key role in converting a variety of messy datasets into clean and structured datasets by creating quality metadata files, running scalable R/Python scripts to model the data and perform data validations. You will also carry data analysis, create data visualizations, and create data models to make sense of data to power critical decisions.



REQUIREMENTS



  • Love R or Python and know how to manipulate data

  • Love when your code throws mind-boggling insights from an almost unusable data

  • Use statistical methods and models to analyze trends in diverse datasets

  • Know when to use bar charts instead of line charts

  • Hate writing the same logic of code twice and love to write scalable, reusable code to process data

  • Create data tools and processes to ease the lives of people working on data processing

  • Not afraid to roll up your sleeves, dig into the code, and implement your ideas


Cookies:



  • You have an updated Kaggle profile and are looking for more challenging problems

  • You have dealt with big data problems


Note: Please do not apply if you haven't worked on a data problem in R or Python Note: If you are looking for advanced machine learning problems and have a good knowledge of ML techniques check out the Machine Learning Intern position

HCL Technologies
  • Dallas, TX

About Our Team

Our team is a mix of data scientists, statistician, economists and data engineers; working on many verticals, including e-Commerce, Supply Chain, Travel & Transportation, Food & Beverages, Insurance, Health Care, Telecom, and many more. We are developing novelty deep learning solutions to solve business challenges that otherwise not be feasible with traditional thinking. We love to code and fail-fast because our projects are mission-critical for our clients.

About Your Role

As a key member of the group, you will have the opportunity to shape our AI-ML strategy:

    • Develop high precision models and tools incorporating state of the art deep learning techniques or leverage battle tested ones
    • Work on projects involving CNNs, RNNs, and other deep reinforcement learning models
    • Build, where applicable, on pre-trained models base on Faster RCNN, SSD, YOLO, etc.
    • Apply expert software development skills to a wide range of ML-related projects
    • Help us monetize our IP
    • Develop and deploy highly distributed cloud services with low-latency, scaling to millions of users
    • Develop algorithms for object detection, natural language understanding, sematic search, etc.
    • Design and develop experiments and analyze their results to build a better service
    • Stay current with state of the art academic research
Your Qualifications
    • MS or PhD in Computer Science or related technical field, or BS with 3+ years of applied experience
    • Possess a real passion for machine learning and statistical methods
    • Our team works closely with many HCL, vendors and client teams to deliver solution, so a history of successful cross-collaboration is a must
    • Strong background in one or more of the following areas: machine learning, natural language processing, computer vision, classifiers, large-scale data mining and data retrieval
    • Experience with one or more ML framework, such as TensorFlow, Keras, Torch, Caffe, Theano, Microsoft Azure ML (big bonus for TensorFlow or Azure ML)
    • Expert coder in Python and Python for ML
    • Experience coder in R, SQL and Java (or C/C++)
    • Extensive experience deploying code to cloud platform (AWS, Azure, GCP) using Docker or Kubernetes containers a major plus
    • Good communication skills at all levels
    • At your best working in a startup-like environment
Preferable
    • Microsoft Chat Bot Framework/ LUIS or IBM Watson/ Conversation Service or Open source NLP Libraries
    • Open source ML libraries and tools like Apache Mahout, Apache Spark
Google
  • San Bruno, CA
Note: By applying to this position your application is automatically submitted to the following locations: Mountain View, CA, USA; New York, NY, USA; San Bruno, CA, USA; Seattle, WA, USA; San Francisco, CA, USA

Minimum qualifications:


  • MS degree in a quantitative discipline (e.g., statistics, operations research, bioinformatics, economics, computational biology, computer science, mathematics, physics, electrical engineering, industrial engineering).

  • 2 years of relevant work experience in data analysis or related field. (e.g., as a statistician / data scientist / computational biologist / bioinformatician).

  • Experience with statistical software (e.g., R, Python, Julia, MATLAB, pandas) and database languages (e.g., SQL).


Preferred qualifications:


  • PhD degree in a quantitative discipline as listed in Minimum Qualifications.

  • 4 years of relevant work experience (e.g., as a statistician / computational biologist / bioinformatician / data scientist), including deep expertise and experience with statistical data analysis such as linear models, multivariate analysis, stochastic models, sampling methods. Analytical engagements outside class work while at school can be included.

  • Applied experience with machine learning on large datasets.

  • Experience articulating business questions and using mathematical techniques to arrive at an answer using available data. Experience translating analysis results into business recommendations.

  • Demonstrated skills in selecting the right statistical tools given a data analysis problem. Demonstrated effective written and verbal communication skills.

  • Demonstrated leadership and self-direction. Demonstrated willingness to both teach others and learn new techniques.

About the job

As a Data Scientist, you will evaluate and improve Google's products. You will collaborate with a multi-disciplinary team of engineers and analysts on a wide range of problems. This position will bring analytical rigor and statistical methods to the challenges of measuring quality, improving consumer products, and understanding the behavior of end-users, advertisers, and publishers.

Google is and always will be an engineering company. We hire people with a broad set of technical skills who are ready to take on some of technology's greatest challenges and make an impact on millions, if not billions, of users. At Google, data scientists not only revolutionize search, they routinely work on massive scalability and storage solutions, large-scale applications and entirely new platforms for developers around the world. From Google Ads to Chrome, Android to YouTube, Social to Local, Google engineers are changing the world one technological achievement after another.

Responsibilities


  • Work with large, complex data sets. Solve difficult, non-routine analysis problems, applying advanced analytical methods as needed. Conduct end-to-end analysis that includes data gathering and requirements specification, processing, analysis, ongoing deliverables, and presentations.

  • Build and prototype analysis pipelines iteratively to provide insights at scale. Develop comprehensive understanding of Google data structures and metrics, advocating for changes where needed for both products development and sales activity.

  • Interact cross-functionally with a wide variety of people and teams. Work closely with engineers to identify opportunities for, design, and assess improvements to google products.

  • Make business recommendations (e.g. cost-benefit, forecasting, experiment analysis) with effective presentations of findings at multiple levels of stakeholders through visual displays of quantitative information.

  • Research and develop analysis, forecasting, and optimization methods to improve the quality of Google's user facing products; example application areas include ads quality, search quality, end-user behavioral modeling, and live experiments.

At Google, we dont just accept differencewe celebrate it, we support it, and we thrive on it for the benefit of our employees, our products and our community. Google is proud to be an equal opportunity workplace and is an affirmative action employer. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity or Veteran status. We also consider qualified applicants regardless of criminal histories, consistent with legal requirements. See also Google's EEO Policy and EEO is the Law. If you have a disability or special need that requires accommodation, please let us know by emailing candidateaccommodations@google.com.
Google
  • New York, NY
Note: By applying to this position your application is automatically submitted to the following locations: Mountain View, CA, USA; New York, NY, USA; San Bruno, CA, USA; Seattle, WA, USA; San Francisco, CA, USA

Minimum qualifications:


  • MS degree in a quantitative discipline (e.g., statistics, operations research, bioinformatics, economics, computational biology, computer science, mathematics, physics, electrical engineering, industrial engineering).

  • 2 years of relevant work experience in data analysis or related field. (e.g., as a statistician / data scientist / computational biologist / bioinformatician).

  • Experience with statistical software (e.g., R, Python, Julia, MATLAB, pandas) and database languages (e.g., SQL).


Preferred qualifications:


  • PhD degree in a quantitative discipline as listed in Minimum Qualifications.

  • 4 years of relevant work experience (e.g., as a statistician / computational biologist / bioinformatician / data scientist), including deep expertise and experience with statistical data analysis such as linear models, multivariate analysis, stochastic models, sampling methods. Analytical engagements outside class work while at school can be included.

  • Applied experience with machine learning on large datasets.

  • Experience articulating business questions and using mathematical techniques to arrive at an answer using available data. Experience translating analysis results into business recommendations.

  • Demonstrated skills in selecting the right statistical tools given a data analysis problem. Demonstrated effective written and verbal communication skills.

  • Demonstrated leadership and self-direction. Demonstrated willingness to both teach others and learn new techniques.

About the job

As a Data Scientist, you will evaluate and improve Google's products. You will collaborate with a multi-disciplinary team of engineers and analysts on a wide range of problems. This position will bring analytical rigor and statistical methods to the challenges of measuring quality, improving consumer products, and understanding the behavior of end-users, advertisers, and publishers.

Google is and always will be an engineering company. We hire people with a broad set of technical skills who are ready to take on some of technology's greatest challenges and make an impact on millions, if not billions, of users. At Google, data scientists not only revolutionize search, they routinely work on massive scalability and storage solutions, large-scale applications and entirely new platforms for developers around the world. From Google Ads to Chrome, Android to YouTube, Social to Local, Google engineers are changing the world one technological achievement after another.

Responsibilities


  • Work with large, complex data sets. Solve difficult, non-routine analysis problems, applying advanced analytical methods as needed. Conduct end-to-end analysis that includes data gathering and requirements specification, processing, analysis, ongoing deliverables, and presentations.

  • Build and prototype analysis pipelines iteratively to provide insights at scale. Develop comprehensive understanding of Google data structures and metrics, advocating for changes where needed for both products development and sales activity.

  • Interact cross-functionally with a wide variety of people and teams. Work closely with engineers to identify opportunities for, design, and assess improvements to google products.

  • Make business recommendations (e.g. cost-benefit, forecasting, experiment analysis) with effective presentations of findings at multiple levels of stakeholders through visual displays of quantitative information.

  • Research and develop analysis, forecasting, and optimization methods to improve the quality of Google's user facing products; example application areas include ads quality, search quality, end-user behavioral modeling, and live experiments.

At Google, we dont just accept differencewe celebrate it, we support it, and we thrive on it for the benefit of our employees, our products and our community. Google is proud to be an equal opportunity workplace and is an affirmative action employer. We are committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity or Veteran status. We also consider qualified applicants regardless of criminal histories, consistent with legal requirements. See also Google's EEO Policy and EEO is the Law. If you have a disability or special need that requires accommodation, please let us know by emailing candidateaccommodations@google.com.
Jefferson Frank
  • Headquarters: Chicago, IL URL
Headquarters: Chicago, IL URL: https://www.jeffersonfrank.com/aws-big-data-jobs [https://www.jeffersonfrank.com/aws-big-data-jobs] AWS Big Data Scientist - Chicago, IL - $90 - $110 per hour Attention Big Data Scientists. I am currently sourcing for a new and exciting new contract position in Chicago. Contract: 6 Months Rate: $90 - $110 Data Scientist Role & Responsibilities: - As a Data Scientist, you will work with our Data Engineers, Machine Learning Engineers, Designers, and Project Managers on interdisciplinary projects - You will work on complex and extremely varied data sets from some of the world largest organizations to solve real world problems - Work within one of the largest and most advanced data science teams - Develop data science products and solutions for clients as well as for our data science team - Write highly optimized code to advance our internal Data Science Toolbox - Add real-world impact to your academic expertise, as you are encouraged to write black papers and present at meetings and conferences should you wish Data Scientist Required Experience & Qualifications: - Experience with Big Data platforms like Spark, and AWS - Experience in statistical modelling and machine learning techniques - Experience in applying data science methods to business problems - Programming experience in at least two of the following languages: R, Python, Scala, SQL - Experience in applying advanced analytical and statistical methods in the commercial world - Good presentation and communication skills, write the ability to explain complex analytical concepts to people from other fields This job will offer the chosen candidate the opportunity to increase their level of experience working with AWS. If you or someone you know are looking for a new opportunity look no further. Please reach out to me either by phone 212-731-8282 (Ext. 3852) or email e.fenco@jeffersonfrank.com [http://mailto:e.fenco@jeffersonfrank.com] Jefferson Frank is the Amazon Web Services (AWS) recruiter of choice. We work with organizations worldwide to find and deliver the best AWS professionals on the planet. Backed by private equity firm TPG Growth, we have a proven track record servicing the AWS permanent and contract recruitment market and, to date, have worked with over 30,000 organizations globally from our offices in North America, Europe, and Asia-Pacific. At Jefferson Frank, our mission is simple: we want happy customers. Whether you're an AWS professional walking into your dream AWS job, or an organization hiring an incredible contractor for your cloud migration project, our goal is to deliver an unrivalled customer experience. Work with us and you'll get the personalized experience you deserve - one you'll simply not find at any other recruitment agency. At Jefferson Frank, we find great people great jobs in AWS. I understand the need for discretion and would welcome the opportunity to speak to any Big Data and cloud analytics candidates that are considering a new career or job either now or in the future. Confidentiality is of the upmost importance. For more information on available AWS Big Data Jobs as well as the cloud market, I can be contacted at 212-731-8282 Ext. 3852. Please see www.jeffersonfrank.com for more information. To apply: https://www.jeffersonfrank.com/aws-big-data-jobs [https://www.jeffersonfrank.com/aws-big-data-jobs]
The Rocket Science Group LLC
  • Atlanta, GA
Mailchimp is a leading marketing platform for small business. We empower millions of customers around the world to build their brands and grow their companies with a suite of marketing automation, multichannel campaign, CRM, and analytics tools.
The Growth team at Mailchimp uses data-driven experimentation to help Mailchimp users get more value out of our products and drive toward achieving our company goals. You will be responsible for producing meaningful analytics to inform the growth process and serve as the subjecting matter expect in all matters related to A/B and MVT test analysis for the growth product team. Youre a skilled collaborator whos able to work across departments and disciplines. You can communicate the value of data-driven experimentation to other cross functional team members. The ideal candidate will have experience in similar marketing optimization, product analytics, or growth analytics role with strong technical and analytical abilities.
As a part of the Growth team, the work you do will have high visibility, as we quickly turn insights into action and drive change on Mailchimp.com and within the Mailchimp product. If this sounds like you, we would love to hear from you!
Responsibilities
    • Translate customer and business needs into actionable analytics that inform Growth strategy and generate test ideas to help meet the goals of the business; Independently act on your recommendations and deliver key insights to the team
    • Partner with qualitative research to work on better understanding the customer journey, provide quantitative insights that help inform the customer journey, and work together to complete analysis and generate test ideas
    • Be a subject matter expert within cross functional Growth Product team, including upholding best practices around A/B testing and educating others on key concepts such as sample size estimation, confidence intervals, and statistical significance
    • Take ideas put forward by the team and create a hypothesis that captures what the team is trying to learn, has clear and measurable KPIs, and can be tested in a reasonable amount of time
    • Provide regular updates and generate quantitative results to show how the Growth team is making progress towards their teams KPI/North Star Metric
    • Create and maintain automated reporting and dashboards to track key marketing and customer experience metrics across multiple properties leveraging SQL, Google BigQuery, and Google Data Studio; monitor for changes in trends, share insights and make recommendations
    • Create requirements for data tracking needs to ensure the hypothesis can be accurately measured and reported on at the end of a test

Requirements
    • Extensive experience in an analytics focused role, product analytics experience is a plus
    • Bachelors or Graduate degree (business or mathematics a plus or equivalent work experience)
    • Desire to work in a fast-paced environment
    • Expertise in A/B testing analytics and best practices, experience with Optimizely is a plus
    • Expertise in SQL, Web Analytics, Excel - R/Python is a plus
    • Proficiency in wrangling and transforming data
    • Strong communication, collaboration, and problem-solving abilities
    • Demonstrated, hands-on experience with data visualization tools
    • Expertise with web analytics tools, specifically Google Analytics (certification preferred)
    • Experience working with Google BigQuery a plus
    • Experience with statistical methods such as regression and hypothesis testing
    • Proven experience analyzing data from a variety of different sources (quantitative and qualitative), presenting the data in a clear and concise manner, and create actionable insights

Mailchimp is a founder-owned and highly profitable company headquartered in the heart of Atlanta. Our purpose is to empower the underdog, and our mission is to democratize cutting edge marketing technology for small business. We offer our employees an exceptional workplace , extremely competitive compensation, fully paid benefits (for employees and their families), and generous profit sharing . We hire humble , collaborative, and ambitious people, and give them endless opportunities to grow and succeed.
We love our hometown and support sustainable urban renewal. Our headquarters is in the historic Ponce City Market , right on the Atlanta Beltline . If you'd like to be considered for this position, please apply below. We look forward to meeting you!
Mailchimp is an equal opportunity employer, and we value diversity at our company. We don't discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.
Comcast
  • Philadelphia, PA

Comcast brings together the best in media and technology. We drive innovation to create the world's best entertainment and online experiences. As a Fortune 50 leader, we set the pace in a variety of innovative and fascinating businesses and create career opportunities across a wide range of locations and disciplines. We are at the forefront of change and move at an amazing pace, thanks to our remarkable people, who bring cutting-edge products and services to life for millions of customers every day. If you share in our passion for teamwork, our vision to revolutionize industries and our goal to lead the future in media and technology, we want you to fast-forward your career at Comcast.

Position Summary:

We are seeking a highly motivated, customer focused; solutions oriented Principal Data Platform Engineer. This position is accountable for developing systems & solutions capable of processing millions of events per second, providing both a real time and historical view into the operation of our wide-array of systems. Design collection and enrichment system components for quality, time lines, scale and reliability. Work on high performance real time data stores and a massive historical data store using best-of-breed and industry leading technology. Build platforms that allow others to design, develop, and apply advanced statistical methods and machine intelligence algorithms.

  • Build core components needed to drive the next generation of data platforms and data processing capability.
  • Build data products, identifying trouble spots, and optimizing the overall user experience with a robust data architecture capable of providing insights
  • Drive standardization and service focused instrumentation
  • Build capabilities that analyze massive amounts of data both in real-time and batch processing
  • Prototype ideas for new tools, products and services across data management ecosystem
  • Employ rigorous continuous delivery practices managed under an agile software development approach
  • Ensure a quality transition to production and solid production operation of the platforms
  • Raise the bar for the Engineering team by advocating leading edge practices such as CI/CD, containerization and TDD
  • Be responsible for end-to-end software development in a DevOps model and enhance DevOps practices to deploy and operate systems
  • Troubleshoot and resolve issues in development, test and production environments
  • Collaborates with project stakeholders across software engineering teams to identify product and technical requirements. Conducts analysis to determine integration needs.
  • Designs new software applications, supports applications under development, and customizes current applications. Assists with the software update process for existing applications, and rollouts of software releases.
  • Researches, writes and edits documentation and technical requirements, including software designs, evaluation plans, test results, technical manuals and formal recommendations and reports.
  • Lead development for new platform initiatives and provide technical leadership throughout the design process and guidance with regards to practices, procedures and techniques. Serves as a guide and mentor for junior-level Software Development Engineers.
  • Works with Quality Assurance team to determine if applications fit specification and technical requirements.
  • Displays in-depth knowledge of, and ability to apply, process design and redesign skills. Presents and defends architectural, design, and technical choices to internal audiences.
  • Multi-task and prioritize with little to no supervision, providing team leadership skills.

Required Skills:

  • Bachelors or Masters in Computer Science, Statistics or related discipline is required.
  • 8+years Data platform & programming experience
  • Experience with Spark(AWS EMR), Spark Streaming &Batch, Hadoop/Cloduera,Hadoop (HDFS, YARN), Hives SQL/ Java, Kafka, MemSQL, Cassandra, HBase, MongoDB (NoSQL), Java, Scala, Git, Maven, Jenkins, Rancher, Puppet, Docker, Kubernetes, Oracle, Teradata and related skills/tools.
  • Experience with Python and R is a plus.
  • Experience in software development of large-scale distributed systems including proven track record of delivering backend systems that participate in a complex ecosystem.
  • Experience in data related technologies and open source frameworks preferred
  • Proficient in Unix/Linux environments
  • Test-driven development/test automation, continuous integration, and deployment automation
  • Enjoy working with data analysis, data quality and reporting
  • Requires understanding of complete SDLC and Experience with continuous integration, test-driven/behavior-driven development, and agile, scrum development methodologies
  • Able to thrive in a fast-paced agile environment and to work effectively across organizational boundaries
  • Excellent communicator, able to analyze and clearly articulate complex issues and technologies understandably and engagingly
  • Great design and problem solving skills
  • Adaptable, proactive and willing to take ownership
  • Keen attention to detail and high level of commitment
  • Data mining, joining, and analyzing big data from multiple environments.
  • Experience with advanced statistical and data mining techniques.
  • Integrating business knowledge with technical functionalities.
  • Providing consultative direction to develop long-term objectives to achieving business objectives.
  • Working with other members of cross-functional team, third party vendors and business partners to deliver quality products in a timely fashion
  • Understanding key business drivers and applying the knowledge of internal/external business issues to improve products and services.
  • Anticipating and interpreting customer needs, assessing requirements and identifying solutions.
  • Ability to initiate and follow through on complex projects of both short and long term duration required.
  • Punctual, regular, and consistent attendance

Comcast is an EOE/Veterans/Disabled/LGBT employer

Expedia, Inc.
  • Bellevue, WA
Are you a curious and passionate about data and how it can be used to solve real world customer problems and drive value? If you’re looking to join a team that is at the forefront of data science and analytics, Expedia Group is laser-focused in nailing the numbers game with access to the best in class techniques and technology. We are looking for extraordinary talent who bring technical expertise, act as business owners, can execute on complex projects and deliver results


What you'll do:



  • Drive customer insights to action as part of a core cross functional team that includes marketing, product and technology

  • Develop techniques to build segments/audiences which can be used for targeting and improved customer profitability

  • Build, assess, and iteratively improve predictive models using best-in-class machine learning methods and best practices within customer lifetime value (CLV) predictions, purchase propensity, traveler preferences and across a variety of other business problems

  • Become the trusted authority on measurement and identification of customer long-term profitability as well as its key drivers

  • Develop a deep understanding of our data infrastructure. Applying that knowledge not only to implement your solutions, but also to champion infrastructure development needs which will help the whole team deliver frictionless, action-orientated customer insight.

  • Be a technical leader on use, adoption and implementation of machine learning and modelling techniques

  • Provide mentorship, peer review, and thought leadership across the business in the areas of statistical methods, experimentation, and tools 


Who you are:



  • Hands on experience in machine learning, building datasets, selecting and engineering features, building and optimizing algorithms. Experience with packages such as R, Python, etc.

  • Strong SQL/data manipulation skills required including cleaning and managing data.

  • Ability to apply a wide variety of statistical methods and data science models to effectively solve business problems

  • Advanced Degree (PhD and/or Master’s) in a quantitative field such as Computer Science, Engineering, Mathematics, Statistics, Machine Learning / AI

  • Knowledge of test design, implementation and evaluation of A/B Testing

  • Real world experience with Internet scale data on a Hadoop or similar platform highly preferred.

  • Experience with data visualization tools and packages a plus.

  • Willingness to work with unstructured, messy data.

  • Self-starter and results orientated, able to work under minimal guidance.

  • Intellectual curiosity and desire to learn new things, techniques and technologies.


Our mission is to revolutionize travel through the power of technology. Collectively, the Expedia Group cover virtually every aspect of researching, planning, and booking travel, from choosing the best airplane seat, to reading personal travel reviews of hotels, to planning what to do in a destination once you arrive. Expedia Group portfolio serves both leisure and business travelers with tastes and budgets ranging from modest to luxury. Expedia delivers consumer travel demand from nearly every continent to thousands of hotels and vacation rentals, hundreds of airlines, thousands of activities, and dozens of car rental companies and cruise lines.


Why join us:
Expedia Group recognizes our success is dependent on the success of our people. We are the world's travel platform, made up of the most knowledgeable, passionate, and creative people in our business. Our brands recognize the power of travel to break down barriers and make people's lives better – that responsibility inspires us to be the place where exceptional people want to do their best work, and to provide them to tools to do so.


Whether you're applying to work in engineering or customer support, marketing or lodging supply, at Expedia Group we act as one team, working towards a common goal; to bring the world within reach. We relentlessly strive for better, but not at the cost of the customer. We act with humility and optimism, respecting ideas big and small. We value diversity and voices of all volumes. We are a global organization but keep our feet on the ground, so we can act fast and stay simple. Our teams also have the chance to give back on a local level and make a difference through our corporate social responsibility program, Expedia Cares. 


If you have a hunger to make a difference with one of the most loved consumer brands in the world and to work in the dynamic travel industry, this is the job for you.

Targa Resources
  • Houston, TX

Job Description

The Data Analyst will take responsibility for managing our master data set, developing reports, and troubleshooting data issues. To do well in this role you need a very fine eye for detail, experience as a data analyst, and deep understanding of the popular data analysis tools and databases. The Data Analyst should be highly skilled in all aspects of data analytics, including mining, generation, and visualization. Additionally, you should be committed to transforming data into readable, goal-driven reports for continued innovation and growth.
 

Tasks / Responsibilities

Develop, implement, and maintain leading-edge analytic systems, taking complicated problems and building simple frameworks  

Identify trends and opportunities for growth through analysis of complex data sets

Use statistical methods to analyze data and generate useful business reports

Work with departmental managers to outline the specific data needs

Evaluate organizational methods and provide source-to-target mappings and information-model specification documents for data sets

Create best-practice reports based on data mining, analysis, and visualization

Evaluate internal systems for efficiency, problems, and inaccuracies, developing and maintaining protocols for handling, processing, and cleaning data

Work directly with management and users to gather requirements, provide status updates, and build relationships
 

Technical Knowledge and Skills

Work closely with business owners to understand and maintain focus on their analytical needs, including identifying critical metrics and KPIs, and deliver actionable insights to relevant decision-makers

Proactively analyze data to answer key questions from stakeholders or out of self-initiated curiosity with an eye for what drives business performance, investigating and communicating areas for improvement in efficiency and productivity

Create and maintain rich interactive visualizations through data interpretation and analysis integrating various reporting components from multiple data sources

Define and implement data acquisition and integration logic, selecting appropriate combination of methods and tools within defined technology stack to ensure optimal scalability and performance of the solution

Develop and maintain databases by acquiring data from primary and secondary sources, and build scripts that will make our data evaluation process more flexible or scalable across data sets

Managing and designing the reporting environment, including data sources, security, and metadata

Supporting data warehouse(s) in identifying and revising reporting requirements

Supporting initiatives for data integrity and normalization

Evaluating changes and updates to source production systems

Providing technical expertise on data storage structures, data mining, and data cleansing

Experience with advanced analytics tools for Object-oriented/object function scripting using languages such as [R, Python, Java, C++, Scala, others]

Experience with popular database programming languages including SQL, PL/SQL, others for relational databases and knowledge and/or certifications on upcoming NoSQL/Hadoop oriented databases like MongoDB, Cassandra, others for nonrelational databases

Knowledge and/or experience in working with SQL on Hadoop tools and technologies including HIVE, Impala, Presto, others from an open source perspective and Hortonworks Data Flow (HDF), Dremio, Informatica, Talend, others from a commercial vendor perspective

Strong experience working with popular data discovery, analytics and BI software tools like Tableau, Qlik, PowerBI and others for semantic-layer-based data discovery

Knowledge and/or familiarity of the midstream services industry and data generated in support of business activities related to the gathering, compressing, treating, processing, and selling natural gas, NGLs and NGL products, and crude oil

Qualifications

Basic Qualifications

A bachelor's degree in computer science, statistics, applied mathematics, data management, information systems, information science or a related quantitative field

3+ years of experience in data mining as a data analyst

Proven analytic skills, including mining, evaluation, analysis, and visualization

Technical writing experience in relevant areas, including queries, reports, and presentations

Other Qualifications

In addition to the basic qualifications listed above, the following other position qualifications are required:

Strong written and verbal communication skills with an aptitude for problem solving

Must be able to independently resolve issues and efficiently self-direct work activities based on the ability to capture, organize, and analyze information

Experience delivering solutions that align with business strategies

Experience troubleshooting complicated issues across multiple systems and driving to solutions

Experience supporting resources performing user acceptance, regression and other required system tests, ensuring that system and/or related integration issues that arise are identified, researched and closed in a timely manner

Experience providing technical solutions to non-technical individuals

Demonstrated team building skills

Ability to deal with internal employees and external business contacts while conveying a positive, service-oriented attitude

ITIL v3 Foundations certified

Willingness to travel to company locations (up to 5%)

Targa offers a very competitive compensation and benefits package.
Targa Resources is an Equal Opportunity Employer.

We are proud to be an EEO/AA employer M/F/D/V.

We maintain a drug-free workplace and perform pre-employment substance abuse testing.

Targa Resources
  • Houston, TX

Job Description

The Data Engineer will be responsible for operationalizing data and analytics initiatives for the company. They will be responsible for expanding and optimizing our data and data pipeline architecture, as well as optimizing data flow and collection. The Data Engineer is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. The Data Engineer will support our software developers, data architects, and data analysts on data initiatives and will ensure optimal data delivery architecture is consistent throughout ongoing projects. They must be self-directed and comfortable supporting the data needs of multiple teams, systems and products.

Tasks / Responsibilities

Develop, construct, test, and maintain data architectures or data pipelines

Ensure data architecture will support the requirements of the business

Discover opportunities for data acquisition

Develop data set processes for data modeling, mining and production

Employ a variety of languages and tools to marry systems together

Recommend ways to improve data reliability, efficiency and quality

Leverage large volumes of data from internal and external sources to answer business demands

Employ sophisticated analytics programs, machine learning and statistical methods to prepare data for use in predictive and prescriptive modeling while exploring and examining data to find hidden patterns

Drive Automation through effective metadata management using innovative and modern tools, techniques and architectures to partially or completely automate the most-common, repeatable and tedious data preparation and integration tasks in order to minimize manual and error-prone processes and improve productivity

Propose appropriate (and innovative) data ingestion, preparation, integration and operationalization techniques in optimally addressing data requirements

Ensure that the data users and consumers use the data provisioned to them responsibly through data governance and compliance initiatives

Promote the available data and analytics capabilities and expertise to business unit leaders and educate them in leveraging these capabilities in achieving their business goals

Technical Knowledge and Skills

Strong experience with advanced analytics tools for Object-oriented/object function scripting using languages such as R, Python, Java, C++, Scala, others

Strong ability to design, build and manage data pipelines for data structures encompassing data transformation, data models, schemas, metadata and workload management

The ability to work with both IT and business in integrating analytics and data science output into business processes and workflows

Strong experience with database programming languages including SQL, PL/SQL, others for relational databases and knowledge and/or certifications on upcoming NoSQL/Hadoop oriented databases like MongoDB, Cassandra, others for nonrelational databases

Strong experience in working with large, heterogeneous datasets in building and optimizing data pipelines, pipeline architectures and integrated datasets using traditional data integration technologies.

Knowledge and/or experience in working with SQL on Hadoop tools and technologies including HIVE, Impala, Presto, others from an open source perspective and Hortonworks Data Flow (HDF), Dremio, Informatica, Talend, others from a commercial vendor perspective

Experience in working with both open-source and commercial message queuing technologies such as Kafka, JMS, Azure Service Bus, Amazon Simple queuing Service, others, stream data integration technologies such as Apache Nifi, Apache Beam, Apache Kafka Streams, Amazon Kinesis, and others

Basic experience working with popular data discovery, analytics and BI software tools like Tableau, Qlik, PowerBI and others for semantic-layer-based data discovery

Strong experience in working with data science teams in refining and optimizing data science and machine learning models and algorithms

Basic experience in working with data governance/data quality and data security teams and specifically data stewards and security resources in moving data pipelines into production with appropriate data quality, governance and security standards and certification

Demonstrated ability to work across multiple deployment environments including cloud, on-premises and hybrid, multiple operating systems and through containerization techniques such as Docker, Kubernetes, AWS Elastic Container Service and others

Familiarity with agile methodologies and capable of applying DevOps and increasingly DataOps principles to data pipelines to improve the communication, integration, reuse and automation of data flows between data managers and consumers across an organization

Knowledge and/or familiarity of the midstream services industry and data generated in support of business activities related to the gathering, compressing, treating, processing, and selling natural gas, NGLs and NGL products, and crude oil


Basic Qualifications

A bachelor's or master's degree in computer science, statistics, applied mathematics, data management, information systems, information science or a related quantitative field or equivalent work experience

At least five years or more of work experience in data management disciplines including data integration, modeling, optimization and data quality, and/or other areas directly relevant to data engineering responsibilities and tasks

At least three years of experience working in cross-functional teams and collaborating with business stakeholders in support of a departmental and/or multi-departmental data management and analytics initiative

Other Qualifications

In addition to the basic qualifications listed above, the following other position qualifications are required:

Strong written and verbal communication skills with an aptitude for problem solving

Must be able to independently resolve issues and efficiently self-direct work activities based on the ability to capture, organize, and analyze information

Experience troubleshooting complicated issues across multiple systems and driving to solutions

Experience providing technical solutions to non-technical individuals

Demonstrated team building skills

Ability to deal with internal employees and external business contacts while conveying a positive, service-oriented attitude

ITIL v3 Foundations certified

Willingness to travel to company locations (up to 5%)


Targa Resources is an Equal Opportunity Employer and does not discriminate against any employee or applicant for employment because of race, color, sex, age, national origin, religion, sexual orientation, gender identity, status as a veteran, and basis of disability or any other federal, state or local protected class.