Description
We are looking for a Data Engineer to help us build and maintain scalable and resilient pipelines that will ingest, process, and deliver the data needed for predictive and descriptive analytics. These data pipelines will further connect to machine learning pipelines to facilitate automatic retraining of our models.
We are a diverse group of data scientists, data engineers, software engineers, machine learning engineers from over 30 different countries. We are smart and fast moving, operating in small teams, with freedom for independent work and fast decision making.
To empower scientists and radically improve how science is published, evaluated and disseminated to researchers, innovators and the public, we have built our own state-of-the-art Artificial Intelligence Review Assistant (AIRA), backed by cutting-edge machine learning algorithms.
Key Responsibilities
Work in a team of machine learning engineers responsible for the productization of prototypes developed by data scientists.
Collaborate with data scientists, machine learning engineers, and other data engineers to design scalable, reliable, and maintainable ETL processes that ensure data scientists and automated ML processes have the necessary data available
Research and adopt the best DataOps & MLOps standards to design and develop scalable end-to-end data pipelines.
Identify opportunities for data process automation.
Establish and enforce best practices (e.g. in development, quality assurance, optimization, release, and monitoring).
Requirements
Degree in Computer Science or similar
Proven experience as a Data Engineer
Proficiency in Python
Experience with a Cloud Platform (e.g. Azure, AWS, GCP)
Experience with a workflow engine (e.g. Data Factory, Airflow)
Experience with SQL and NoSQL (e.g. MongoDB) databases
Experience with Hadoop & Spark
Great communication, teamwork, problem-solving, and organizational skills.
Nice To Have
Understanding of supervised and unsupervised machine learning algorithms
Stream-processing frameworks (e.g. Kafka)
Benefits
Competitive salary.
Participation in Frontiers annual bonus scheme
25 leave days + 4 well-being days (pro rata and expiring each year on 31st of December)
Great work-life balance.
Opportunity to work remotely
Fresh fruit, snacks and coffee.
English classes.
Team building/sport activities and monthly social events.
Lots of opportunities to work with exciting technologies and solve challenging problems
Who we are
Frontiers is an award-winning open science platform and leading open access scholarly publisher. We are one of the largest and most cited publishers globally. Our journals span science, health, humanities and social sciences, engineering, and sustainability and we continue to expand into new academic disciplines so more researchers can publish open access.
Dec 23, 2021
Full time
Description
We are looking for a Data Engineer to help us build and maintain scalable and resilient pipelines that will ingest, process, and deliver the data needed for predictive and descriptive analytics. These data pipelines will further connect to machine learning pipelines to facilitate automatic retraining of our models.
We are a diverse group of data scientists, data engineers, software engineers, machine learning engineers from over 30 different countries. We are smart and fast moving, operating in small teams, with freedom for independent work and fast decision making.
To empower scientists and radically improve how science is published, evaluated and disseminated to researchers, innovators and the public, we have built our own state-of-the-art Artificial Intelligence Review Assistant (AIRA), backed by cutting-edge machine learning algorithms.
Key Responsibilities
Work in a team of machine learning engineers responsible for the productization of prototypes developed by data scientists.
Collaborate with data scientists, machine learning engineers, and other data engineers to design scalable, reliable, and maintainable ETL processes that ensure data scientists and automated ML processes have the necessary data available
Research and adopt the best DataOps & MLOps standards to design and develop scalable end-to-end data pipelines.
Identify opportunities for data process automation.
Establish and enforce best practices (e.g. in development, quality assurance, optimization, release, and monitoring).
Requirements
Degree in Computer Science or similar
Proven experience as a Data Engineer
Proficiency in Python
Experience with a Cloud Platform (e.g. Azure, AWS, GCP)
Experience with a workflow engine (e.g. Data Factory, Airflow)
Experience with SQL and NoSQL (e.g. MongoDB) databases
Experience with Hadoop & Spark
Great communication, teamwork, problem-solving, and organizational skills.
Nice To Have
Understanding of supervised and unsupervised machine learning algorithms
Stream-processing frameworks (e.g. Kafka)
Benefits
Competitive salary.
Participation in Frontiers annual bonus scheme
25 leave days + 4 well-being days (pro rata and expiring each year on 31st of December)
Great work-life balance.
Opportunity to work remotely
Fresh fruit, snacks and coffee.
English classes.
Team building/sport activities and monthly social events.
Lots of opportunities to work with exciting technologies and solve challenging problems
Who we are
Frontiers is an award-winning open science platform and leading open access scholarly publisher. We are one of the largest and most cited publishers globally. Our journals span science, health, humanities and social sciences, engineering, and sustainability and we continue to expand into new academic disciplines so more researchers can publish open access.
JOB DESCRIPTION You're ready to gain the skills and experience needed to grow within your role and advance your career - and we have the perfect software engineering opportunity for you. As a Software Engineer II Python at JPMorgan Chase within the Capital Technology team, you are part of an agile team that works to enhance, design, and deliver the software components of the firm's state-of-the-art technology products in a secure, stable, and scalable way. As an emerging member of a software engineering team, you execute software solutions through the design, development, and technical troubleshooting of multiple components within a technical product, application, or system, while gaining the skills and experience needed to grow within your role. Job responsibilities Executes standard software solutions, design, development, and technical troubleshooting Writes secure and high-quality code using the syntax of at least one programming language with limited guidance Designs, develops, codes, and troubleshoots with consideration of upstream and downstream systems and technical implications Applies knowledge of tools within the Software Development Life Cycle toolchain to improve the value realized by automation Applies technical troubleshooting to break down solutions and solve technical problems of basic complexity Gathers, analyzes, and draws conclusions from large, diverse data sets to identify problems and contribute to decision-making in service of secure, stable application development Learns and applies system processes, methodologies, and skills for the development of secure, stable code and systems Adds to team culture of diversity, equity, inclusion, and respect Required qualifications, capabilities, and skills Formal training or certification on software engineering concepts and applied experience. Skilled with Python or PySpark Hands-on practical experience in system design, application development, testing, and operational stability Experience in developing, debugging, and maintaining code in a large corporate environment with one or more modern programming languages and database querying languages Demonstrable ability to code in one or more languages Experience across the whole Software Development Life Cycle Exposure to agile methodologies such as CI/CD, Application Resiliency, and Security Emerging knowledge of software applications and technical processes within a technical discipline (e.g., cloud, artificial intelligence, machine learning, mobile, etc.) Preferred qualifications, capabilities, and skills Exposure to cloud technologies (Airflow, Astronomer,Databricks Kubernetes, AWS, Spark, Kafka) Experience in Financial Service Industry is a bonus. ABOUT US J.P. Morgan is a global leader in financial services, providing strategic advice and products to the world's most prominent corporations, governments, wealthy individuals and institutional investors. Our first-class business in a first-class way approach to serving clients drives everything we do. We strive to build trusted, long-term partnerships to help our clients achieve their business objectives. We recognize that our people are our strength and the diverse talents they bring to our global workforce are directly linked to our success. We are an equal opportunity employer and place a high value on diversity and inclusion at our company. We do not discriminate on the basis of any protected attribute, including race, religion, color, national origin, gender, sexual orientation, gender identity, gender expression, age, marital or veteran status, pregnancy or disability, or any other basis protected under applicable law. We also make reasonable accommodations for applicants' and employees' religious practices and beliefs, as well as mental health or physical disability needs. Visit our FAQs for more information about requesting an accommodation. ABOUT THE TEAM Our professionals in our Corporate Functions cover a diverse range of areas from finance and risk to human resources and marketing. Our corporate teams are an essential part of our company, ensuring that we're setting our businesses, clients, customers and employees up for success.
Apr 17, 2024
Full time
JOB DESCRIPTION You're ready to gain the skills and experience needed to grow within your role and advance your career - and we have the perfect software engineering opportunity for you. As a Software Engineer II Python at JPMorgan Chase within the Capital Technology team, you are part of an agile team that works to enhance, design, and deliver the software components of the firm's state-of-the-art technology products in a secure, stable, and scalable way. As an emerging member of a software engineering team, you execute software solutions through the design, development, and technical troubleshooting of multiple components within a technical product, application, or system, while gaining the skills and experience needed to grow within your role. Job responsibilities Executes standard software solutions, design, development, and technical troubleshooting Writes secure and high-quality code using the syntax of at least one programming language with limited guidance Designs, develops, codes, and troubleshoots with consideration of upstream and downstream systems and technical implications Applies knowledge of tools within the Software Development Life Cycle toolchain to improve the value realized by automation Applies technical troubleshooting to break down solutions and solve technical problems of basic complexity Gathers, analyzes, and draws conclusions from large, diverse data sets to identify problems and contribute to decision-making in service of secure, stable application development Learns and applies system processes, methodologies, and skills for the development of secure, stable code and systems Adds to team culture of diversity, equity, inclusion, and respect Required qualifications, capabilities, and skills Formal training or certification on software engineering concepts and applied experience. Skilled with Python or PySpark Hands-on practical experience in system design, application development, testing, and operational stability Experience in developing, debugging, and maintaining code in a large corporate environment with one or more modern programming languages and database querying languages Demonstrable ability to code in one or more languages Experience across the whole Software Development Life Cycle Exposure to agile methodologies such as CI/CD, Application Resiliency, and Security Emerging knowledge of software applications and technical processes within a technical discipline (e.g., cloud, artificial intelligence, machine learning, mobile, etc.) Preferred qualifications, capabilities, and skills Exposure to cloud technologies (Airflow, Astronomer,Databricks Kubernetes, AWS, Spark, Kafka) Experience in Financial Service Industry is a bonus. ABOUT US J.P. Morgan is a global leader in financial services, providing strategic advice and products to the world's most prominent corporations, governments, wealthy individuals and institutional investors. Our first-class business in a first-class way approach to serving clients drives everything we do. We strive to build trusted, long-term partnerships to help our clients achieve their business objectives. We recognize that our people are our strength and the diverse talents they bring to our global workforce are directly linked to our success. We are an equal opportunity employer and place a high value on diversity and inclusion at our company. We do not discriminate on the basis of any protected attribute, including race, religion, color, national origin, gender, sexual orientation, gender identity, gender expression, age, marital or veteran status, pregnancy or disability, or any other basis protected under applicable law. We also make reasonable accommodations for applicants' and employees' religious practices and beliefs, as well as mental health or physical disability needs. Visit our FAQs for more information about requesting an accommodation. ABOUT THE TEAM Our professionals in our Corporate Functions cover a diverse range of areas from finance and risk to human resources and marketing. Our corporate teams are an essential part of our company, ensuring that we're setting our businesses, clients, customers and employees up for success.
JOB DESCRIPTION We have an exciting and rewarding opportunity for you to take your software engineering career to the next level. As a Python / Spark Big Data Software Engineer III at JPMorgan Chase within the Capital Reporting product, you serve as a seasoned member of an agile team to design and deliver trusted market-leading technology products in a secure, stable, and scalable way. You are responsible for carrying out critical technology solutions across multiple technical areas within various business functions in support of the firm's business objectives. Job responsibilities : Executes software solutions, design, development, and technical troubleshooting with ability to think beyond routine or conventional approaches to build solutions or break down technical problems Produces architecture and design artifacts for complex applications while being accountable for ensuring design constraints are met by software code development Proactively identifies hidden problems and patterns in data and uses these insights to drive improvements to coding hygiene and system architecture Contributes to software engineering communities of practice and events that explore new and emerging technologies Adds to team culture of diversity, equity, inclusion, and respect Required qualifications, capabilities, and skills : Formal training or certification on cloud or microservice architecture concepts and proficient applied experience Demonstrated knowledge of software applications and technical processes within a cloud or microservice architecture. Hands-on practical experience in system design, application development, testing, and operational stability Experience in developing, debugging, and maintaining code in a large corporate environment with one or more modern programming languages and database querying languages Overall knowledge of the Software Development Life Cycle Solid understanding of agile methodologies such as CI/CD, Applicant Resiliency, and Security Preferred qualifications, capabilities, and skills : Skilled with Python or PySpark Exposure to cloud technologies Databricks,Airflow, Astronomer, Kubernetes, AWS, Spark, Kafka) Experience with Big Data solutions or Relational DB. Experience in Financial Service Industry is a bonus. ABOUT US J.P. Morgan is a global leader in financial services, providing strategic advice and products to the world's most prominent corporations, governments, wealthy individuals and institutional investors. Our first-class business in a first-class way approach to serving clients drives everything we do. We strive to build trusted, long-term partnerships to help our clients achieve their business objectives. We recognize that our people are our strength and the diverse talents they bring to our global workforce are directly linked to our success. We are an equal opportunity employer and place a high value on diversity and inclusion at our company. We do not discriminate on the basis of any protected attribute, including race, religion, color, national origin, gender, sexual orientation, gender identity, gender expression, age, marital or veteran status, pregnancy or disability, or any other basis protected under applicable law. We also make reasonable accommodations for applicants' and employees' religious practices and beliefs, as well as mental health or physical disability needs. Visit our FAQs for more information about requesting an accommodation. ABOUT THE TEAM Our professionals in our Corporate Functions cover a diverse range of areas from finance and risk to human resources and marketing. Our corporate teams are an essential part of our company, ensuring that we're setting our businesses, clients, customers and employees up for success.
Apr 17, 2024
Full time
JOB DESCRIPTION We have an exciting and rewarding opportunity for you to take your software engineering career to the next level. As a Python / Spark Big Data Software Engineer III at JPMorgan Chase within the Capital Reporting product, you serve as a seasoned member of an agile team to design and deliver trusted market-leading technology products in a secure, stable, and scalable way. You are responsible for carrying out critical technology solutions across multiple technical areas within various business functions in support of the firm's business objectives. Job responsibilities : Executes software solutions, design, development, and technical troubleshooting with ability to think beyond routine or conventional approaches to build solutions or break down technical problems Produces architecture and design artifacts for complex applications while being accountable for ensuring design constraints are met by software code development Proactively identifies hidden problems and patterns in data and uses these insights to drive improvements to coding hygiene and system architecture Contributes to software engineering communities of practice and events that explore new and emerging technologies Adds to team culture of diversity, equity, inclusion, and respect Required qualifications, capabilities, and skills : Formal training or certification on cloud or microservice architecture concepts and proficient applied experience Demonstrated knowledge of software applications and technical processes within a cloud or microservice architecture. Hands-on practical experience in system design, application development, testing, and operational stability Experience in developing, debugging, and maintaining code in a large corporate environment with one or more modern programming languages and database querying languages Overall knowledge of the Software Development Life Cycle Solid understanding of agile methodologies such as CI/CD, Applicant Resiliency, and Security Preferred qualifications, capabilities, and skills : Skilled with Python or PySpark Exposure to cloud technologies Databricks,Airflow, Astronomer, Kubernetes, AWS, Spark, Kafka) Experience with Big Data solutions or Relational DB. Experience in Financial Service Industry is a bonus. ABOUT US J.P. Morgan is a global leader in financial services, providing strategic advice and products to the world's most prominent corporations, governments, wealthy individuals and institutional investors. Our first-class business in a first-class way approach to serving clients drives everything we do. We strive to build trusted, long-term partnerships to help our clients achieve their business objectives. We recognize that our people are our strength and the diverse talents they bring to our global workforce are directly linked to our success. We are an equal opportunity employer and place a high value on diversity and inclusion at our company. We do not discriminate on the basis of any protected attribute, including race, religion, color, national origin, gender, sexual orientation, gender identity, gender expression, age, marital or veteran status, pregnancy or disability, or any other basis protected under applicable law. We also make reasonable accommodations for applicants' and employees' religious practices and beliefs, as well as mental health or physical disability needs. Visit our FAQs for more information about requesting an accommodation. ABOUT THE TEAM Our professionals in our Corporate Functions cover a diverse range of areas from finance and risk to human resources and marketing. Our corporate teams are an essential part of our company, ensuring that we're setting our businesses, clients, customers and employees up for success.
JOB DESCRIPTION If you are looking for a game-changing career, working for one of the world's leading financial institutions, you've come to the right place. As a Principal Software Engineer at JPMorgan Chase within Global Finance Technology, you provide expertise and engineering excellence as an integral part of an agile team to enhance, build, and deliver trusted market-leading technology products in a secure, stable, and scalable way. Leverage your advanced technical capabilities and collaborate with colleagues across the organization to drive best-in-class outcomes across various technologies to support one or more of the firm's portfolios. Job responsibilities Creates complex and scalable coding frameworks using appropriate software design frameworks Develops secure and high-quality production code, and reviews and debugs code written by others Advises cross-functional teams on technological matters within your domain of expertise Serves as the function's go-to subject matter expert Contributes to the development of technical methods in specialized fields in line with the latest product development methodologies Creates durable, reusable software frameworks that are leveraged across teams and functions Influences leaders and senior stakeholders across business, product, and technology teams Champions the firm's culture of diversity, equity, inclusion, and respect Required qualifications, capabilities, and skills Formal training or certification on software engineering concepts and expert applied experience Extensive development experience using Java/Scala Demonstrated experience building distributed computing applications using Big Data Frameworks like Hadoop, Spark and/or Databricks Deep experience using and coding on big data frameworks like Hadoop, Spark and/or Kafka Hands-on practical experience delivering system design, application development, testing, and operational stability Expertise in object oriented and/or functional programming language(s) Understanding of Public Cloud technologies and experience building cloud native applications Advanced knowledge of software architecture, application development, and technical processes with considerable in-depth knowledge in one or more technical disciplines (e.g., cloud, artificial intelligence, machine learning, mobile, etc.) Ability to present and effectively communicate to Senior Business and Technology Executives along with Engineers at all levels Preferred qualifications, capabilities, and skills Demonstrated experience with cloud technologies like AWS, GCP or Azure Domain knowledge of the financial services industry with focus on building data pipelines for financial reporting Expertise in software engineering practices like TDD, ATDD, BDD, Pair Programming ABOUT US JPMorgan Chase & Co., one of the oldest financial institutions, offers innovative financial solutions to millions of consumers, small businesses and many of the world's most prominent corporate, institutional and government clients under the J.P. Morgan and Chase brands. Our history spans over 200 years and today we are a leader in investment banking, consumer and small business banking, commercial banking, financial transaction processing and asset management. We offer a competitive total rewards package including base salary determined based on the role, experience, skill set, and location. For those in eligible roles, we offer discretionary incentive compensation which may be awarded in recognition of firm performance and individual achievements and contributions. We also offer a range of benefits and programs to meet employee needs, based on eligibility. These benefits include comprehensive health care coverage, on-site health and wellness centers, a retirement savings plan, backup childcare, tuition reimbursement, mental health support, financial coaching and more. Additional details about total compensation and benefits will be provided during the hiring process. We recognize that our people are our strength and the diverse talents they bring to our global workforce are directly linked to our success. We are an equal opportunity employer and place a high value on diversity and inclusion at our company. We do not discriminate on the basis of any protected attribute, including race, religion, color, national origin, gender, sexual orientation, gender identity, gender expression, age, marital or veteran status, pregnancy or disability, or any other basis protected under applicable law. We also make reasonable accommodations for applicants' and employees' religious practices and beliefs, as well as mental health or physical disability needs. Visit our FAQs for more information about requesting an accommodation. JPMorgan Chase is an Equal Opportunity Employer, including Disability/Veterans ABOUT THE TEAM Our professionals in our Corporate Functions cover a diverse range of areas from finance and risk to human resources and marketing. Our corporate teams are an essential part of our company, ensuring that we're setting our businesses, clients, customers and employees up for success.
Apr 17, 2024
Full time
JOB DESCRIPTION If you are looking for a game-changing career, working for one of the world's leading financial institutions, you've come to the right place. As a Principal Software Engineer at JPMorgan Chase within Global Finance Technology, you provide expertise and engineering excellence as an integral part of an agile team to enhance, build, and deliver trusted market-leading technology products in a secure, stable, and scalable way. Leverage your advanced technical capabilities and collaborate with colleagues across the organization to drive best-in-class outcomes across various technologies to support one or more of the firm's portfolios. Job responsibilities Creates complex and scalable coding frameworks using appropriate software design frameworks Develops secure and high-quality production code, and reviews and debugs code written by others Advises cross-functional teams on technological matters within your domain of expertise Serves as the function's go-to subject matter expert Contributes to the development of technical methods in specialized fields in line with the latest product development methodologies Creates durable, reusable software frameworks that are leveraged across teams and functions Influences leaders and senior stakeholders across business, product, and technology teams Champions the firm's culture of diversity, equity, inclusion, and respect Required qualifications, capabilities, and skills Formal training or certification on software engineering concepts and expert applied experience Extensive development experience using Java/Scala Demonstrated experience building distributed computing applications using Big Data Frameworks like Hadoop, Spark and/or Databricks Deep experience using and coding on big data frameworks like Hadoop, Spark and/or Kafka Hands-on practical experience delivering system design, application development, testing, and operational stability Expertise in object oriented and/or functional programming language(s) Understanding of Public Cloud technologies and experience building cloud native applications Advanced knowledge of software architecture, application development, and technical processes with considerable in-depth knowledge in one or more technical disciplines (e.g., cloud, artificial intelligence, machine learning, mobile, etc.) Ability to present and effectively communicate to Senior Business and Technology Executives along with Engineers at all levels Preferred qualifications, capabilities, and skills Demonstrated experience with cloud technologies like AWS, GCP or Azure Domain knowledge of the financial services industry with focus on building data pipelines for financial reporting Expertise in software engineering practices like TDD, ATDD, BDD, Pair Programming ABOUT US JPMorgan Chase & Co., one of the oldest financial institutions, offers innovative financial solutions to millions of consumers, small businesses and many of the world's most prominent corporate, institutional and government clients under the J.P. Morgan and Chase brands. Our history spans over 200 years and today we are a leader in investment banking, consumer and small business banking, commercial banking, financial transaction processing and asset management. We offer a competitive total rewards package including base salary determined based on the role, experience, skill set, and location. For those in eligible roles, we offer discretionary incentive compensation which may be awarded in recognition of firm performance and individual achievements and contributions. We also offer a range of benefits and programs to meet employee needs, based on eligibility. These benefits include comprehensive health care coverage, on-site health and wellness centers, a retirement savings plan, backup childcare, tuition reimbursement, mental health support, financial coaching and more. Additional details about total compensation and benefits will be provided during the hiring process. We recognize that our people are our strength and the diverse talents they bring to our global workforce are directly linked to our success. We are an equal opportunity employer and place a high value on diversity and inclusion at our company. We do not discriminate on the basis of any protected attribute, including race, religion, color, national origin, gender, sexual orientation, gender identity, gender expression, age, marital or veteran status, pregnancy or disability, or any other basis protected under applicable law. We also make reasonable accommodations for applicants' and employees' religious practices and beliefs, as well as mental health or physical disability needs. Visit our FAQs for more information about requesting an accommodation. JPMorgan Chase is an Equal Opportunity Employer, including Disability/Veterans ABOUT THE TEAM Our professionals in our Corporate Functions cover a diverse range of areas from finance and risk to human resources and marketing. Our corporate teams are an essential part of our company, ensuring that we're setting our businesses, clients, customers and employees up for success.
At Kpler, we simplify global trade information and provide valuable insights. Founded in 2014, our goal is to help over 10,000 organisations by offering the best intelligence on commodities, energy, and maritime through a single platform. Working at Kpler means you'll be a key player in turning complex data into strategic resources for our clients. Your role involves creating data-driven stories that empower clients in their industries. Your expertise helps Kpler navigate markets successfully. Your journey starts here, where innovation meets impact. Join our team of 500+ talented people from 35+ countries worldwide. Purpose of the role: Maritime domain awareness involves the effective fusion of fragmented pieces of information that compose the complex maritime landscape. Vessel tracking data and trading history, satellite imagery, sanction lists and OSINT comprise some of the data sources which, when combined efficiently, provide an understanding of events that take place at ports and in open seas. Rapidly evolving global geopolitical developments and tensions greatly affect trading, disturbing the supply chain and altering the patterns of shipping activity. As a senior data scientist with Kpler's risk and compliance team you will develop algorithms that will detect deviations from normal maritime activity and will discriminate trivial abnormalities from those that imply deceptive shipping practices. This venture requires dealing with high volumes of dispersed and often highly imbalanced data. Some of the exciting projects you will work on include: Driving the design, rapid experimentation, development, testing and deployment of models for maritime anomaly detection and risk assessment. Optimising and fine-tuning models in production, overseeing the continuous monitoring of deployed models, effectively handling model and data drifts in a timely manner. Building robust pipelines for integrating data from diverse sources, including big geospatial data, mobility data, noisy sensor data, tabular data. Researching and identifying methods, data sources, features that will drive business impact in the scope of maritime risk and compliance. As a senior data scientist, you will: Devise efficient solutions to tackle ML/Big Data challenges using relevant and up-to-date methods and technologies. Work across the stack to deliver new features end to end. Ensure optimal, cost-effective design decisions that improve performance and overcome scalability limits. Own meaningful parts of our service, demonstrating the ability to lead projects independently, have an impact, and grow with the company. Identify opportunities for novel projects and liaise with product teams to advance ideas into value-adding features. Actively share knowledge and document insights, effectively communicate complex concepts and analyses to technical and non-technical audiences, aiming to support continuous team improvement and drive collaboration. Act as a mentor for our junior data scientists, helping to accelerate their growth; you will act as the Tech Lead on some projects. Be part of a vibrant Machine Learning community in Kpler, tackling the whole spectrum of ML problems. Our Machine Learning tech stack includes: Python, ML libraries (TensorFlow, pytorch, scikit-learn, transformers, XGBoost, ResNet), geospatial libraries (shapely, geopandas, rasterio), CV libraries (scikit-image, OpenCV, yolo, Detectron2). AWS, Postgres, Apache Airflow, Apache kafka, Apache Spark. Mandatory requirements: You have at least 5 years of experience in DS role, deploying models into production; You have proven experience delivering end-to-end ML solutions that produced business value. You have proven experience with big data technologies, specifically Spark and Kafka You are proficient in Python. You have expert knowledge of at least one cloud computing platform (preferably aws). You are fluent in English. Nice to haves but not mandatory: You have expertise on applications with focus on geospatial data and mobility analytics (highly desirable). You have a Ph. D. in a quantitative field (computer science, mathematics, physics, engineering ). You are familiar with the shipping industry and/or commodity trading. You are comfortable with software engineering best practices. You value code simplicity, performance and attention to detail. You have experience working in an international environment. We're a dynamic company dedicated to nurturing connections and innovating solutions that tackle market challenges head-on. If you're driven by customer satisfaction and thrive on turning ideas into reality, then you've found your ideal destination. Are you prepared to embark on this exciting journey with us? we make things happen We act decisively and with purpose, and we like to go the extra mile. we build together We foster relationships and develop creative solutions to address market challenges with cool features and solutions. hey, how can i help you today? Being accessible and supportive to colleagues and clients with a friendly approach is essential. Our People Pledge Don't meet every single requirement? Research shows that women and people of color are less likely than others to apply if they feel like they don't match 100% of the job requirements. Don't let the confidence gap stand in your way, we'd love to hear from you! We understand that experience comes in many different forms and are dedicated to adding new perspectives to the team. Kpler is committed to providing a fair, inclusive and diverse work-environment. We believe that different perspectives lead to better ideas, and better ideas allow us to better understand the needs and interests of our diverse, global community. We welcome people of different backgrounds, experiences, abilities and perspectives and are an equal opportunity employer. By applying, I confirm that I have read and accept the Staff Privacy Notice
Apr 16, 2024
Full time
At Kpler, we simplify global trade information and provide valuable insights. Founded in 2014, our goal is to help over 10,000 organisations by offering the best intelligence on commodities, energy, and maritime through a single platform. Working at Kpler means you'll be a key player in turning complex data into strategic resources for our clients. Your role involves creating data-driven stories that empower clients in their industries. Your expertise helps Kpler navigate markets successfully. Your journey starts here, where innovation meets impact. Join our team of 500+ talented people from 35+ countries worldwide. Purpose of the role: Maritime domain awareness involves the effective fusion of fragmented pieces of information that compose the complex maritime landscape. Vessel tracking data and trading history, satellite imagery, sanction lists and OSINT comprise some of the data sources which, when combined efficiently, provide an understanding of events that take place at ports and in open seas. Rapidly evolving global geopolitical developments and tensions greatly affect trading, disturbing the supply chain and altering the patterns of shipping activity. As a senior data scientist with Kpler's risk and compliance team you will develop algorithms that will detect deviations from normal maritime activity and will discriminate trivial abnormalities from those that imply deceptive shipping practices. This venture requires dealing with high volumes of dispersed and often highly imbalanced data. Some of the exciting projects you will work on include: Driving the design, rapid experimentation, development, testing and deployment of models for maritime anomaly detection and risk assessment. Optimising and fine-tuning models in production, overseeing the continuous monitoring of deployed models, effectively handling model and data drifts in a timely manner. Building robust pipelines for integrating data from diverse sources, including big geospatial data, mobility data, noisy sensor data, tabular data. Researching and identifying methods, data sources, features that will drive business impact in the scope of maritime risk and compliance. As a senior data scientist, you will: Devise efficient solutions to tackle ML/Big Data challenges using relevant and up-to-date methods and technologies. Work across the stack to deliver new features end to end. Ensure optimal, cost-effective design decisions that improve performance and overcome scalability limits. Own meaningful parts of our service, demonstrating the ability to lead projects independently, have an impact, and grow with the company. Identify opportunities for novel projects and liaise with product teams to advance ideas into value-adding features. Actively share knowledge and document insights, effectively communicate complex concepts and analyses to technical and non-technical audiences, aiming to support continuous team improvement and drive collaboration. Act as a mentor for our junior data scientists, helping to accelerate their growth; you will act as the Tech Lead on some projects. Be part of a vibrant Machine Learning community in Kpler, tackling the whole spectrum of ML problems. Our Machine Learning tech stack includes: Python, ML libraries (TensorFlow, pytorch, scikit-learn, transformers, XGBoost, ResNet), geospatial libraries (shapely, geopandas, rasterio), CV libraries (scikit-image, OpenCV, yolo, Detectron2). AWS, Postgres, Apache Airflow, Apache kafka, Apache Spark. Mandatory requirements: You have at least 5 years of experience in DS role, deploying models into production; You have proven experience delivering end-to-end ML solutions that produced business value. You have proven experience with big data technologies, specifically Spark and Kafka You are proficient in Python. You have expert knowledge of at least one cloud computing platform (preferably aws). You are fluent in English. Nice to haves but not mandatory: You have expertise on applications with focus on geospatial data and mobility analytics (highly desirable). You have a Ph. D. in a quantitative field (computer science, mathematics, physics, engineering ). You are familiar with the shipping industry and/or commodity trading. You are comfortable with software engineering best practices. You value code simplicity, performance and attention to detail. You have experience working in an international environment. We're a dynamic company dedicated to nurturing connections and innovating solutions that tackle market challenges head-on. If you're driven by customer satisfaction and thrive on turning ideas into reality, then you've found your ideal destination. Are you prepared to embark on this exciting journey with us? we make things happen We act decisively and with purpose, and we like to go the extra mile. we build together We foster relationships and develop creative solutions to address market challenges with cool features and solutions. hey, how can i help you today? Being accessible and supportive to colleagues and clients with a friendly approach is essential. Our People Pledge Don't meet every single requirement? Research shows that women and people of color are less likely than others to apply if they feel like they don't match 100% of the job requirements. Don't let the confidence gap stand in your way, we'd love to hear from you! We understand that experience comes in many different forms and are dedicated to adding new perspectives to the team. Kpler is committed to providing a fair, inclusive and diverse work-environment. We believe that different perspectives lead to better ideas, and better ideas allow us to better understand the needs and interests of our diverse, global community. We welcome people of different backgrounds, experiences, abilities and perspectives and are an equal opportunity employer. By applying, I confirm that I have read and accept the Staff Privacy Notice
If you need assistance during the recruiting process due to a disability, please reach out to our Recruiting Accommodations Team through the Accommodation Request form . This form is used only by individuals with disabilities who require assistance or adjustments in applying and interviewing for a job. This form is not for inquiring about a position or the status of an application. Are you interested in harnessing machine learning to power new personalisation experiences for travellers? Do you believe that modern online experiences should adapt to the needs and preferences of individual users? Are you fascinated by data, machine learning techniques, and software systems engineering? Do you love building creative, high-scale distributed systems using a diverse set of state of the art technologies? Our team wants to talk to you! What you'll do You work in a cross-functional team of ML Engineers and ML Scientists to operationalise ML models in production. You build scalable, high-performance systems for model development, data ingestion, feature engineering, inference, and monitoring/evaluation. You provide an accurate time estimates for your scope of work, turn it into code, and deliver on schedule. You advocate for quality coding. Write secure, stable, testable, maintainable code with minimal defects. Who you are 2+ years experience in ML and software engineering for Bachelor's, 1+ years for Master's Developed software in a team environment of at least 5 engineers (agile, version control, etc.). Built and maintained an ML model or pipeline in production environments in public/hybrid cloud infrastructure. Coding proficiency in at least one modern programming language (Java, Scala, Python etc.). Strong background in data structures and algorithms Experience working with at least one machine learning framework (TensorFlow, PyTorch, XGBoost, etc) Experience working with big data technologies (Spark, Kafka, Hive, Databricks, feature stores, etc). Experience working with containerisation, deployment and orchestration technologies (Docker, Kubernetes, Airflow, CI/CD pipelines, etc) Experience with automated testing, including unit, functional, and integration testing Excellent organizational and communication skills Bachelor's or master's degree in CS or similar About Expedia Group Expedia Group (NASDAQ: EXPE) powers travel for everyone, everywhere through our global platform. Driven by the core belief that travel is a force for good, we help people experience the world in new ways and build lasting connections. We provide industry-leading technology solutions to fuel partner growth and success, while facilitating memorable experiences for travelers. Expedia Group's family of brands includes: Brand Expedia, Expedia Partner Solutions, Vrbo, trivago, Orbitz, Travelocity, Hotwire, Wotif, ebookers, CheapTickets, Expedia Group Media Solutions, Expedia Local Expert, and Expedia Cruises. Employment opportunities and job offers at Expedia Group will always come from Expedia Group's Talent Acquisition and hiring teams. Never provide sensitive, personal information to someone unless you're confident who the recipient is. Expedia Group does not extend job offers via email or any other messaging tools to individuals to whom we have not made prior contact. Our email domain The official website to find and apply for job openings at Expedia Group is . Expedia is committed to creating an inclusive work environment with a diverse workforce. All qualified applicants will receive consideration for employment without regard to race, religion, gender, sexual orientation, national origin, disability or age. If you need assistance during the recruiting process due to a disability, please reach out to our Recruiting Accommodations Team through the Accommodation Request form . This form is used only by individuals with disabilities who require assistance or adjustments in applying and interviewing for a job. This form is not for inquiring about a position or the status of an application. Machine Learning Engineer II Are you interested in harnessing machine learning to power new personalisation experiences for travellers? Do you believe that modern online experiences should adapt to the needs and preferences of individual users? Are you fascinated by data, machine learning techniques, and software systems engineering? Do you love building creative, high-scale distributed systems using a diverse set of state of the art technologies? Our team wants to talk to you! What you'll do You work in a cross-functional team of ML Engineers and ML Scientists to operationalise ML models in production. You build scalable, high-performance systems for model development, data ingestion, feature engineering, inference, and monitoring/evaluation. You provide an accurate time estimates for your scope of work, turn it into code, and deliver on schedule. You advocate for quality coding. Write secure, stable, testable, maintainable code with minimal defects. Who you are 2+ years experience in ML and software engineering for Bachelor's, 1+ years for Master's Developed software in a team environment of at least 5 engineers (agile, version control, etc.). Built and maintained an ML model or pipeline in production environments in public/hybrid cloud infrastructure. Coding proficiency in at least one modern programming language (Java, Scala, Python etc.). Strong background in data structures and algorithms Experience working with at least one machine learning framework (TensorFlow, PyTorch, XGBoost, etc) Experience working with big data technologies (Spark, Kafka, Hive, Databricks, feature stores, etc). Experience working with containerisation, deployment and orchestration technologies (Docker, Kubernetes, Airflow, CI/CD pipelines, etc) Experience with automated testing, including unit, functional, and integration testing Excellent organizational and communication skills Bachelor's or master's degree in CS or similar About Expedia Group Expedia Group (NASDAQ: EXPE) powers travel for everyone, everywhere through our global platform. Driven by the core belief that travel is a force for good, we help people experience the world in new ways and build lasting connections. We provide industry-leading technology solutions to fuel partner growth and success, while facilitating memorable experiences for travelers. Expedia Group's family of brands includes: Brand Expedia, Expedia Partner Solutions, Vrbo, trivago, Orbitz, Travelocity, Hotwire, Wotif, ebookers, CheapTickets, Expedia Group Media Solutions, Expedia Local Expert, and Expedia Cruises. 2021 Expedia, Inc. All rights reserved. Trademarks and logos are the property of their respective owners. CST: -50 Employment opportunities and job offers at Expedia Group will always come from Expedia Group's Talent Acquisition and hiring teams. Never provide sensitive, personal information to someone unless you're confident who the recipient is. Expedia Group does not extend job offers via email or any other messaging tools to individuals to whom we have not made prior contact. Our email domain The official website to find and apply for job openings at Expedia Group is . Expedia is committed to creating an inclusive work environment with a diverse workforce. All qualified applicants will receive consideration for employment without regard to race, religion, gender, sexual orientation, national origin, disability or age. Growth Marketing Manager, SEO and Data Insights At Expedia Group, we're committed to providing an inclusive and accessible recruiting experience for candidates with disabilities. If you require an accommodation or adjustment for any part of the application or hiring process, please let us know by completing our Accommodation Request form.
Apr 16, 2024
Full time
If you need assistance during the recruiting process due to a disability, please reach out to our Recruiting Accommodations Team through the Accommodation Request form . This form is used only by individuals with disabilities who require assistance or adjustments in applying and interviewing for a job. This form is not for inquiring about a position or the status of an application. Are you interested in harnessing machine learning to power new personalisation experiences for travellers? Do you believe that modern online experiences should adapt to the needs and preferences of individual users? Are you fascinated by data, machine learning techniques, and software systems engineering? Do you love building creative, high-scale distributed systems using a diverse set of state of the art technologies? Our team wants to talk to you! What you'll do You work in a cross-functional team of ML Engineers and ML Scientists to operationalise ML models in production. You build scalable, high-performance systems for model development, data ingestion, feature engineering, inference, and monitoring/evaluation. You provide an accurate time estimates for your scope of work, turn it into code, and deliver on schedule. You advocate for quality coding. Write secure, stable, testable, maintainable code with minimal defects. Who you are 2+ years experience in ML and software engineering for Bachelor's, 1+ years for Master's Developed software in a team environment of at least 5 engineers (agile, version control, etc.). Built and maintained an ML model or pipeline in production environments in public/hybrid cloud infrastructure. Coding proficiency in at least one modern programming language (Java, Scala, Python etc.). Strong background in data structures and algorithms Experience working with at least one machine learning framework (TensorFlow, PyTorch, XGBoost, etc) Experience working with big data technologies (Spark, Kafka, Hive, Databricks, feature stores, etc). Experience working with containerisation, deployment and orchestration technologies (Docker, Kubernetes, Airflow, CI/CD pipelines, etc) Experience with automated testing, including unit, functional, and integration testing Excellent organizational and communication skills Bachelor's or master's degree in CS or similar About Expedia Group Expedia Group (NASDAQ: EXPE) powers travel for everyone, everywhere through our global platform. Driven by the core belief that travel is a force for good, we help people experience the world in new ways and build lasting connections. We provide industry-leading technology solutions to fuel partner growth and success, while facilitating memorable experiences for travelers. Expedia Group's family of brands includes: Brand Expedia, Expedia Partner Solutions, Vrbo, trivago, Orbitz, Travelocity, Hotwire, Wotif, ebookers, CheapTickets, Expedia Group Media Solutions, Expedia Local Expert, and Expedia Cruises. Employment opportunities and job offers at Expedia Group will always come from Expedia Group's Talent Acquisition and hiring teams. Never provide sensitive, personal information to someone unless you're confident who the recipient is. Expedia Group does not extend job offers via email or any other messaging tools to individuals to whom we have not made prior contact. Our email domain The official website to find and apply for job openings at Expedia Group is . Expedia is committed to creating an inclusive work environment with a diverse workforce. All qualified applicants will receive consideration for employment without regard to race, religion, gender, sexual orientation, national origin, disability or age. If you need assistance during the recruiting process due to a disability, please reach out to our Recruiting Accommodations Team through the Accommodation Request form . This form is used only by individuals with disabilities who require assistance or adjustments in applying and interviewing for a job. This form is not for inquiring about a position or the status of an application. Machine Learning Engineer II Are you interested in harnessing machine learning to power new personalisation experiences for travellers? Do you believe that modern online experiences should adapt to the needs and preferences of individual users? Are you fascinated by data, machine learning techniques, and software systems engineering? Do you love building creative, high-scale distributed systems using a diverse set of state of the art technologies? Our team wants to talk to you! What you'll do You work in a cross-functional team of ML Engineers and ML Scientists to operationalise ML models in production. You build scalable, high-performance systems for model development, data ingestion, feature engineering, inference, and monitoring/evaluation. You provide an accurate time estimates for your scope of work, turn it into code, and deliver on schedule. You advocate for quality coding. Write secure, stable, testable, maintainable code with minimal defects. Who you are 2+ years experience in ML and software engineering for Bachelor's, 1+ years for Master's Developed software in a team environment of at least 5 engineers (agile, version control, etc.). Built and maintained an ML model or pipeline in production environments in public/hybrid cloud infrastructure. Coding proficiency in at least one modern programming language (Java, Scala, Python etc.). Strong background in data structures and algorithms Experience working with at least one machine learning framework (TensorFlow, PyTorch, XGBoost, etc) Experience working with big data technologies (Spark, Kafka, Hive, Databricks, feature stores, etc). Experience working with containerisation, deployment and orchestration technologies (Docker, Kubernetes, Airflow, CI/CD pipelines, etc) Experience with automated testing, including unit, functional, and integration testing Excellent organizational and communication skills Bachelor's or master's degree in CS or similar About Expedia Group Expedia Group (NASDAQ: EXPE) powers travel for everyone, everywhere through our global platform. Driven by the core belief that travel is a force for good, we help people experience the world in new ways and build lasting connections. We provide industry-leading technology solutions to fuel partner growth and success, while facilitating memorable experiences for travelers. Expedia Group's family of brands includes: Brand Expedia, Expedia Partner Solutions, Vrbo, trivago, Orbitz, Travelocity, Hotwire, Wotif, ebookers, CheapTickets, Expedia Group Media Solutions, Expedia Local Expert, and Expedia Cruises. 2021 Expedia, Inc. All rights reserved. Trademarks and logos are the property of their respective owners. CST: -50 Employment opportunities and job offers at Expedia Group will always come from Expedia Group's Talent Acquisition and hiring teams. Never provide sensitive, personal information to someone unless you're confident who the recipient is. Expedia Group does not extend job offers via email or any other messaging tools to individuals to whom we have not made prior contact. Our email domain The official website to find and apply for job openings at Expedia Group is . Expedia is committed to creating an inclusive work environment with a diverse workforce. All qualified applicants will receive consideration for employment without regard to race, religion, gender, sexual orientation, national origin, disability or age. Growth Marketing Manager, SEO and Data Insights At Expedia Group, we're committed to providing an inclusive and accessible recruiting experience for candidates with disabilities. If you require an accommodation or adjustment for any part of the application or hiring process, please let us know by completing our Accommodation Request form.
At Kpler, we simplify global trade information and provide valuable insights. Founded in 2014, our goal is to help over 10,000 organisations by offering the best intelligence on commodities, energy, and maritime through a single platform. Working at Kpler means you'll be a key player in turning complex data into strategic resources for our clients. Your role involves creating data-driven stories that empower clients in their industries. Your expertise helps Kpler navigate markets successfully. Your journey starts here, where innovation meets impact. Join our team of 500+ talented people from 35+ countries worldwide. Purpose of the rol Maritime domain awareness and commodity tracking involve the effective fusion of fragmented information that compose the complex commodity flows landscape. When combined efficiently, vessel tracking data and trading history, commodity supply and demand, proprietary trading data, and machine learning algorithms provide an understanding of events at ports and on open seas at different levels of granularity. Rapidly evolving global geopolitical developments and tensions significantly affect trading, disturbing the supply chain and altering commodity flow patterns which brings more uncertainty about the modelling of these phenomena. As a senior data scientist with Kpler's Flows team, you will develop algorithms that extend our comprehension of commodities flows in space and time by devising forecasting models, predicting future trades, detect anomalies in data and anomalous behaviour of various players. Some of the exciting projects you will work on include: Driving the design, rapid experimentation, development, testing and deployment of data science models for flow forecast models and anomaly detection; Optimising and fine-tuning models in production, overseeing the continuous monitoring of deployed models, and effectively handling model and data drift promptly; Building robust pipelines for integrating data from diverse sources, including big geospatial data, ship mobility data, and document recognition; Researching and identifying methods, data sources, and features that will drive business impact and improve models' accuracy in the current scope of ever-changing world-scale commodity trading; As a senior data scientist, you will: Devise efficient solutions to tackle ML/Big Data challenges using relevant, up-to-date methods and technologies. Work across the stack to deliver new features end-to-end, from prototyping, to deployment and caring for data drift in production. Ensure optimal, cost-effective design decisions that improve performance and overcome scalability limits. Own meaningful parts of our service, demonstrating the ability to lead projects independently, have an impact, and grow with the company. Identify opportunities for novel projects and liaise with product teams to advance ideas into value-adding features. Actively share knowledge and document insights, effectively communicate complex concepts and analysis to technical and non-technical audiences, aiming to support continuous team improvement and drive collaboration. Act as a mentor for our junior data scientists, helping to accelerate their growth; you will act as the Tech Lead on some projects. Be part of a vibrant Machine Learning community in Kpler, tackling the whole spectrum of ML problems. Our Machine Learning tech stack includes: Python, ML libraries (TensorFlow, pytorch, scikit-learn, transformers, XGBoost, ResNet), geospatial libraries (shapely, geopandas, rasterio), CV libraries (scikit-image, OpenCV, yolo, Detectron2). AWS, Postgres, Apache Airflow, Apache kafka, Apache Spark Mandatory requirements: You have at least 5 years of experience in the DS role, deploying models into production; You have proven experience delivering end-to-end ML solutions that produce business value. You are proficient in Python. You have expert knowledge of at least one cloud computing platform (preferably AWS). You are fluent in English. Nice to haves but not mandatory: You have expertise on applications focusing on geospatial data and mobility analytics (highly desirable). You have proven experience with big data technologies, specifically Spark and Kafka. You have experience working with state-of-art ML pipeline technologies (such as MLflow, Sagemaker ) or building a ML pipeline by yourself (Docker, Kubernetes, Paperspace, Airflow ). You have a Ph. D. in a quantitative field (computer science, mathematics, physics, engineering ). You are familiar with the shipping industry and commodity trading. You are comfortable with software engineering best practices. You value code simplicity, performance and attention to detail. You have experience working in an international environment. We're a dynamic company dedicated to nurturing connections and innovating solutions that tackle market challenges head-on. If you're driven by customer satisfaction and thrive on turning ideas into reality, then you've found your ideal destination. Are you prepared to embark on this exciting journey with us? we make things happen We act decisively and with purpose, and we like to go the extra mile. we build together We foster relationships and develop creative solutions to address market challenges with cool features and solutions. hey, how can i help you today? Being accessible and supportive to colleagues and clients with a friendly approach is essential. Our People Pledge Don't meet every single requirement? Research shows that women and people of color are less likely than others to apply if they feel like they don't match 100% of the job requirements. Don't let the confidence gap stand in your way, we'd love to hear from you! We understand that experience comes in many different forms and are dedicated to adding new perspectives to the team. Kpler is committed to providing a fair, inclusive and diverse work-environment. We believe that different perspectives lead to better ideas, and better ideas allow us to better understand the needs and interests of our diverse, global community. We welcome people of different backgrounds, experiences, abilities and perspectives and are an equal opportunity employer. By applying, I confirm that I have read and accept the Staff Privacy Notice
Apr 12, 2024
Full time
At Kpler, we simplify global trade information and provide valuable insights. Founded in 2014, our goal is to help over 10,000 organisations by offering the best intelligence on commodities, energy, and maritime through a single platform. Working at Kpler means you'll be a key player in turning complex data into strategic resources for our clients. Your role involves creating data-driven stories that empower clients in their industries. Your expertise helps Kpler navigate markets successfully. Your journey starts here, where innovation meets impact. Join our team of 500+ talented people from 35+ countries worldwide. Purpose of the rol Maritime domain awareness and commodity tracking involve the effective fusion of fragmented information that compose the complex commodity flows landscape. When combined efficiently, vessel tracking data and trading history, commodity supply and demand, proprietary trading data, and machine learning algorithms provide an understanding of events at ports and on open seas at different levels of granularity. Rapidly evolving global geopolitical developments and tensions significantly affect trading, disturbing the supply chain and altering commodity flow patterns which brings more uncertainty about the modelling of these phenomena. As a senior data scientist with Kpler's Flows team, you will develop algorithms that extend our comprehension of commodities flows in space and time by devising forecasting models, predicting future trades, detect anomalies in data and anomalous behaviour of various players. Some of the exciting projects you will work on include: Driving the design, rapid experimentation, development, testing and deployment of data science models for flow forecast models and anomaly detection; Optimising and fine-tuning models in production, overseeing the continuous monitoring of deployed models, and effectively handling model and data drift promptly; Building robust pipelines for integrating data from diverse sources, including big geospatial data, ship mobility data, and document recognition; Researching and identifying methods, data sources, and features that will drive business impact and improve models' accuracy in the current scope of ever-changing world-scale commodity trading; As a senior data scientist, you will: Devise efficient solutions to tackle ML/Big Data challenges using relevant, up-to-date methods and technologies. Work across the stack to deliver new features end-to-end, from prototyping, to deployment and caring for data drift in production. Ensure optimal, cost-effective design decisions that improve performance and overcome scalability limits. Own meaningful parts of our service, demonstrating the ability to lead projects independently, have an impact, and grow with the company. Identify opportunities for novel projects and liaise with product teams to advance ideas into value-adding features. Actively share knowledge and document insights, effectively communicate complex concepts and analysis to technical and non-technical audiences, aiming to support continuous team improvement and drive collaboration. Act as a mentor for our junior data scientists, helping to accelerate their growth; you will act as the Tech Lead on some projects. Be part of a vibrant Machine Learning community in Kpler, tackling the whole spectrum of ML problems. Our Machine Learning tech stack includes: Python, ML libraries (TensorFlow, pytorch, scikit-learn, transformers, XGBoost, ResNet), geospatial libraries (shapely, geopandas, rasterio), CV libraries (scikit-image, OpenCV, yolo, Detectron2). AWS, Postgres, Apache Airflow, Apache kafka, Apache Spark Mandatory requirements: You have at least 5 years of experience in the DS role, deploying models into production; You have proven experience delivering end-to-end ML solutions that produce business value. You are proficient in Python. You have expert knowledge of at least one cloud computing platform (preferably AWS). You are fluent in English. Nice to haves but not mandatory: You have expertise on applications focusing on geospatial data and mobility analytics (highly desirable). You have proven experience with big data technologies, specifically Spark and Kafka. You have experience working with state-of-art ML pipeline technologies (such as MLflow, Sagemaker ) or building a ML pipeline by yourself (Docker, Kubernetes, Paperspace, Airflow ). You have a Ph. D. in a quantitative field (computer science, mathematics, physics, engineering ). You are familiar with the shipping industry and commodity trading. You are comfortable with software engineering best practices. You value code simplicity, performance and attention to detail. You have experience working in an international environment. We're a dynamic company dedicated to nurturing connections and innovating solutions that tackle market challenges head-on. If you're driven by customer satisfaction and thrive on turning ideas into reality, then you've found your ideal destination. Are you prepared to embark on this exciting journey with us? we make things happen We act decisively and with purpose, and we like to go the extra mile. we build together We foster relationships and develop creative solutions to address market challenges with cool features and solutions. hey, how can i help you today? Being accessible and supportive to colleagues and clients with a friendly approach is essential. Our People Pledge Don't meet every single requirement? Research shows that women and people of color are less likely than others to apply if they feel like they don't match 100% of the job requirements. Don't let the confidence gap stand in your way, we'd love to hear from you! We understand that experience comes in many different forms and are dedicated to adding new perspectives to the team. Kpler is committed to providing a fair, inclusive and diverse work-environment. We believe that different perspectives lead to better ideas, and better ideas allow us to better understand the needs and interests of our diverse, global community. We welcome people of different backgrounds, experiences, abilities and perspectives and are an equal opportunity employer. By applying, I confirm that I have read and accept the Staff Privacy Notice
AWS Data Engineer - Insight & Data Services - Permanent Salary guideline: 75,000 - 95,000 pa (D.O.E) + 5-10% Bonus, Pension up to 6% Contributory, Health Insurance, Life Assurance etc. Base Location: Closest office to your home location / Hybrid working / Part Remote / UK-wide The Client: Our client's Insights and Data practice is the leading Data Science and AI Engineering provider in the United Kingdom with over 450 consultants serving the UK market. They are the true market leader! The Role: As an AWS Data Engineer within the Insights & Data Emerging Tech Team, this role is a unique chance to make a real difference in your career and to make a difference that affects people's lives and transforms the way companies and governments operate. Do you want to amaze people, to take them on a journey and show them something truly fantastic? Do you want to be at the forefront of the AI revolution? The Focus of the Role: We are looking for strong AWS Data Engineers who are passionate about Cloud technology. Your work will be to: Design and build data engineering solutions and support the planning and implementation of data platform services including sizing, configuration, and needs assessment Build relationships with client stakeholders to establish a high level of rapport and confidence Work with clients, local teams and offshore resources to deliver modern data products Work effectively on client sites, Capgemini offices and from home Use AWS Data focused Reference Architecture Design and build data service APIs Analyse current business practices, processes and procedures and identify future opportunities for leveraging AWS services Implement effective metrics and monitoring processes Essential Skills & Experienced needed: Have a deep, hands-on design and engineering background in AWS, across a wide range of AWS services with the ability to demonstrate working on large engagements Experience of AWS tools (e.g. Athena, Redshift, Glue, EMR) Java, Scala, Python, Spark, SQL Experience of developing enterprise grade ETL/ELT data pipelines. Deep understanding of data manipulation/wrangling techniques Demonstrable knowledge of applying Data Engineering best practices (coding practices to DS, unit testing, version control, code review). Big Data Eco-Systems, Cloudera/Hortonworks, AWS EMR, GCP DataProc or GCP Cloud Data Fusion. NoSQL Databases. Dynamo DB/Neo4j/Elastic, Google Cloud Datastore. Snowflake Data Warehouse/Platform Streaming technologies and processing engines, Kinesis, Kafka, Pub/Sub and Spark Streaming. Experience of working with CI/CD technologies, Git, Jenkins, Spinnaker, GCP Cloud Build, Ansible etc Experience building and deploying solutions to Cloud (AWS, Google Cloud) including Cloud provisioning tools Have hands on experience with Infrastructure-as-Code technologies: Terraform, Ansible Capable of working in either an agile or Waterfall development environment, both as part of a team and individually E2E Solution Design skills - Prototyping, Usability testing Experience with SQL and NoSQL modern data stores. Strong interpersonal skills with the ability to work with clients to establish requirements in non-technical language. Ability to translate business requirements into plausible technical solutions for articulation to other development staff. Good understanding of Data Governance, including Master Data Management (MDM) and Data Quality tools and processes Influencing and supporting project delivery through involvement in project/sprint planning and QA Nice to have - "Desirable": Knowledge of other cloud platforms Google Data Products tools knowledge (e.g., BigQuery, Dataflow, DataProc, AI Building Blocks, Looker, Cloud Data Fusion, Data prep, etc.) Relevant certifications Python Snowflake Databricks To apply please click the "Apply" button and follow the instructions. For a further discussion, please contact Sam Stark - (phone number removed) 83DATA is a boutique consultancy specialising in Data Engineering and Architecture Data Science (ML, AI, DL) Data Visualisation RPA within the UK. We provide high-quality interim and permanent senior IT professionals.
Apr 10, 2024
Full time
AWS Data Engineer - Insight & Data Services - Permanent Salary guideline: 75,000 - 95,000 pa (D.O.E) + 5-10% Bonus, Pension up to 6% Contributory, Health Insurance, Life Assurance etc. Base Location: Closest office to your home location / Hybrid working / Part Remote / UK-wide The Client: Our client's Insights and Data practice is the leading Data Science and AI Engineering provider in the United Kingdom with over 450 consultants serving the UK market. They are the true market leader! The Role: As an AWS Data Engineer within the Insights & Data Emerging Tech Team, this role is a unique chance to make a real difference in your career and to make a difference that affects people's lives and transforms the way companies and governments operate. Do you want to amaze people, to take them on a journey and show them something truly fantastic? Do you want to be at the forefront of the AI revolution? The Focus of the Role: We are looking for strong AWS Data Engineers who are passionate about Cloud technology. Your work will be to: Design and build data engineering solutions and support the planning and implementation of data platform services including sizing, configuration, and needs assessment Build relationships with client stakeholders to establish a high level of rapport and confidence Work with clients, local teams and offshore resources to deliver modern data products Work effectively on client sites, Capgemini offices and from home Use AWS Data focused Reference Architecture Design and build data service APIs Analyse current business practices, processes and procedures and identify future opportunities for leveraging AWS services Implement effective metrics and monitoring processes Essential Skills & Experienced needed: Have a deep, hands-on design and engineering background in AWS, across a wide range of AWS services with the ability to demonstrate working on large engagements Experience of AWS tools (e.g. Athena, Redshift, Glue, EMR) Java, Scala, Python, Spark, SQL Experience of developing enterprise grade ETL/ELT data pipelines. Deep understanding of data manipulation/wrangling techniques Demonstrable knowledge of applying Data Engineering best practices (coding practices to DS, unit testing, version control, code review). Big Data Eco-Systems, Cloudera/Hortonworks, AWS EMR, GCP DataProc or GCP Cloud Data Fusion. NoSQL Databases. Dynamo DB/Neo4j/Elastic, Google Cloud Datastore. Snowflake Data Warehouse/Platform Streaming technologies and processing engines, Kinesis, Kafka, Pub/Sub and Spark Streaming. Experience of working with CI/CD technologies, Git, Jenkins, Spinnaker, GCP Cloud Build, Ansible etc Experience building and deploying solutions to Cloud (AWS, Google Cloud) including Cloud provisioning tools Have hands on experience with Infrastructure-as-Code technologies: Terraform, Ansible Capable of working in either an agile or Waterfall development environment, both as part of a team and individually E2E Solution Design skills - Prototyping, Usability testing Experience with SQL and NoSQL modern data stores. Strong interpersonal skills with the ability to work with clients to establish requirements in non-technical language. Ability to translate business requirements into plausible technical solutions for articulation to other development staff. Good understanding of Data Governance, including Master Data Management (MDM) and Data Quality tools and processes Influencing and supporting project delivery through involvement in project/sprint planning and QA Nice to have - "Desirable": Knowledge of other cloud platforms Google Data Products tools knowledge (e.g., BigQuery, Dataflow, DataProc, AI Building Blocks, Looker, Cloud Data Fusion, Data prep, etc.) Relevant certifications Python Snowflake Databricks To apply please click the "Apply" button and follow the instructions. For a further discussion, please contact Sam Stark - (phone number removed) 83DATA is a boutique consultancy specialising in Data Engineering and Architecture Data Science (ML, AI, DL) Data Visualisation RPA within the UK. We provide high-quality interim and permanent senior IT professionals.
Site Name: USA - Pennsylvania - Upper Providence, UK - Hertfordshire - Stevenage, UK - London - Brentford, USA - Pennsylvania - Philadelphia Posted Date: Oct The mission of the Data Science and Data Engineering (DSDE) organization within GSK Pharmaceuticals R&D is to get the right data, to the right people, at the right time. TheData Framework and Opsorganization ensures we can do this efficiently, reliably, transparently, and at scale through the creation of a leading-edge, cloud-native data services framework. We focus heavily on developer experience, on strong, semantic abstractions for the data ecosystem, on professional operations and aggressive automation, and on transparency of operations and cost. We are looking for a skilled Data Framework Engineer II to join our growing team. The Data Framework team builds and manages (in partnership with Tech) reusable components and architectures designed to make it both fast and easy to build robust, scalable, production-grade data products and services in the challenging biomedical data space. A Data Framework Engineer IIknows the metrics desired for their tools andservices anditerates to deliver and improve on those metrics in an agile fashion. A Data Framework Engineer II is a highly technical individual contributor, building modern, cloud-native systems for standardizing and templatizing data engineering, such as: Standardized physical storage and search / indexing systems Schema management (data + metadata + versioning + provenance + governance) API semantics and ontology management Standard API architectures Kafka + standard streaming semantics Standard components for publishing data to file-based, relational, and other sorts of data stores Metadata systems Tooling for QA / evaluation Additional responsibilities also include: Given a well-specified data framework problem, implement end-to-end solutionsusing appropriate programming languages(e.g.Python,Scala, or Go), open-source tools (e.g.Spark,Elasticsearch, ...), and cloud vendor-provided tools (e.g.Amazon S3) Leverage tools provided by Tech (e.g.infrastructure as code, CloudOps,DevOps, logging / alerting, ...) in delivery ofsolutions Write proper documentation in code as well as in wikis/other documentationsystems Writefantastic code along withthe proper unit, functional, and integration tests for code and services to ensurequality Stayup to datewith developments in theopen-sourcecommunity around data engineering, data science, and similartooling The DSDE team is built on the principles of ownership, accountability, continuous development, and collaboration. We hire for the long term, and we're motivated to make this a great place to work. Our leaders will be committed to your career and development from day one. Why you? Basic Qualifications: We are looking for professionals with these required skills to achieve our goals: PhD in Computer Science with a focus in Data Engineering, DataOps, DevOps, MLOps, Software Engineering OR Masters and 2+ years experience Experience with common distributed data tools (Spark, Kafka, etc) Experience with basics of data architecture (e.g. optimizing physical layout for access patterns) Experience with basics of search engines/indexing (e.g. Elasticsearch, Lucene) Demonstrated experience in writing Python, Scala, Go, and/or C++ Preferred Qualifications: If you have the following characteristics, it would be a plus: Experience with agile software development Experience building and designing a DevOps-first way of working Demonstrated experience building reusable components on top of the CNCF ecosystem including Kubernetes (or similar ecosystem) Experience with schema tools and schema management (Avro, Protobuf) Why GSK? Our values and expectations are at the heart of everything we do and form an important part of our culture. These include Patient focus, Transparency, Respect, Integrity along with Courage, Accountability, Development, and Teamwork. As GSK focuses on our values and expectations and a culture of innovation, performance, and trust, the successful candidate will demonstrate the following capabilities: Operating at pace and agile decision making - using evidence and applying judgement to balance pace, rigour and risk. Committed to delivering high-quality results, overcoming challenges, focusing on what matters, execution. Continuously looking for opportunities to learn, build skills and share learning. Sustaining energy and wellbeing Building strong relationships and collaboration, honest and open conversations. Budgeting and cost consciousness LI-GSK If you require an accommodation or other assistance to apply for a job at GSK, please contact the GSK Service Centre at 1- (US Toll Free) or +1 (outside US). GSK is an Equal Opportunity Employer and, in the US, we adhere to Affirmative Action principles. This ensures that all qualified applicants will receive equal consideration for employment without regard to race, color, national origin, religion, sex, pregnancy, marital status, sexual orientation, gender identity/expression, age, disability, genetic information, military service, covered/protected veteran status or any other federal, state or local protected class. At GSK, the health and safety of our employees are of paramount importance. As a science-led healthcare company on a mission to get ahead of disease together, we believe that supporting vaccination against COVID-19 is the single best thing we can do in the US to ensure the health and safety of our employees, complementary workers, workplaces, customers, consumers, communities, and the patients we serve. GSK has made the decision to require all US employees to be fully vaccinated against COVID-19, where allowed by state or local law and where vaccine supply is readily available. The only exceptions to this requirement are employees who are approved for an accommodation for religious, medical or disability-related reasons. Important notice to Employment businesses/ Agencies GSK does not accept referrals from employment businesses and/or employment agencies in respect of the vacancies posted on this site. All employment businesses/agencies are required to contact GSK's commercial and general procurement/human resources department to obtain prior written authorization before referring any candidates to GSK. The obtaining of prior written authorization is a condition precedent to any agreement (verbal or written) between the employment business/ agency and GSK. In the absence of such written authorization being obtained any actions undertaken by the employment business/agency shall be deemed to have been performed without the consent or contractual agreement of GSK. GSK shall therefore not be liable for any fees arising from such actions or any fees arising from any referrals by employment businesses/agencies in respect of the vacancies posted on this site. Please note that if you are a US Licensed Healthcare Professional or Healthcare Professional as defined by the laws of the state issuing your license, GSK may be required to capture and report expenses GSK incurs, on your behalf, in the event you are afforded an interview for employment. This capture of applicable transfers of value is necessary to ensure GSK's compliance to all federal and state US Transparency requirements. For more information, please visit GSK's Transparency Reporting For the Record site.
Sep 24, 2022
Full time
Site Name: USA - Pennsylvania - Upper Providence, UK - Hertfordshire - Stevenage, UK - London - Brentford, USA - Pennsylvania - Philadelphia Posted Date: Oct The mission of the Data Science and Data Engineering (DSDE) organization within GSK Pharmaceuticals R&D is to get the right data, to the right people, at the right time. TheData Framework and Opsorganization ensures we can do this efficiently, reliably, transparently, and at scale through the creation of a leading-edge, cloud-native data services framework. We focus heavily on developer experience, on strong, semantic abstractions for the data ecosystem, on professional operations and aggressive automation, and on transparency of operations and cost. We are looking for a skilled Data Framework Engineer II to join our growing team. The Data Framework team builds and manages (in partnership with Tech) reusable components and architectures designed to make it both fast and easy to build robust, scalable, production-grade data products and services in the challenging biomedical data space. A Data Framework Engineer IIknows the metrics desired for their tools andservices anditerates to deliver and improve on those metrics in an agile fashion. A Data Framework Engineer II is a highly technical individual contributor, building modern, cloud-native systems for standardizing and templatizing data engineering, such as: Standardized physical storage and search / indexing systems Schema management (data + metadata + versioning + provenance + governance) API semantics and ontology management Standard API architectures Kafka + standard streaming semantics Standard components for publishing data to file-based, relational, and other sorts of data stores Metadata systems Tooling for QA / evaluation Additional responsibilities also include: Given a well-specified data framework problem, implement end-to-end solutionsusing appropriate programming languages(e.g.Python,Scala, or Go), open-source tools (e.g.Spark,Elasticsearch, ...), and cloud vendor-provided tools (e.g.Amazon S3) Leverage tools provided by Tech (e.g.infrastructure as code, CloudOps,DevOps, logging / alerting, ...) in delivery ofsolutions Write proper documentation in code as well as in wikis/other documentationsystems Writefantastic code along withthe proper unit, functional, and integration tests for code and services to ensurequality Stayup to datewith developments in theopen-sourcecommunity around data engineering, data science, and similartooling The DSDE team is built on the principles of ownership, accountability, continuous development, and collaboration. We hire for the long term, and we're motivated to make this a great place to work. Our leaders will be committed to your career and development from day one. Why you? Basic Qualifications: We are looking for professionals with these required skills to achieve our goals: PhD in Computer Science with a focus in Data Engineering, DataOps, DevOps, MLOps, Software Engineering OR Masters and 2+ years experience Experience with common distributed data tools (Spark, Kafka, etc) Experience with basics of data architecture (e.g. optimizing physical layout for access patterns) Experience with basics of search engines/indexing (e.g. Elasticsearch, Lucene) Demonstrated experience in writing Python, Scala, Go, and/or C++ Preferred Qualifications: If you have the following characteristics, it would be a plus: Experience with agile software development Experience building and designing a DevOps-first way of working Demonstrated experience building reusable components on top of the CNCF ecosystem including Kubernetes (or similar ecosystem) Experience with schema tools and schema management (Avro, Protobuf) Why GSK? Our values and expectations are at the heart of everything we do and form an important part of our culture. These include Patient focus, Transparency, Respect, Integrity along with Courage, Accountability, Development, and Teamwork. As GSK focuses on our values and expectations and a culture of innovation, performance, and trust, the successful candidate will demonstrate the following capabilities: Operating at pace and agile decision making - using evidence and applying judgement to balance pace, rigour and risk. Committed to delivering high-quality results, overcoming challenges, focusing on what matters, execution. Continuously looking for opportunities to learn, build skills and share learning. Sustaining energy and wellbeing Building strong relationships and collaboration, honest and open conversations. Budgeting and cost consciousness LI-GSK If you require an accommodation or other assistance to apply for a job at GSK, please contact the GSK Service Centre at 1- (US Toll Free) or +1 (outside US). GSK is an Equal Opportunity Employer and, in the US, we adhere to Affirmative Action principles. This ensures that all qualified applicants will receive equal consideration for employment without regard to race, color, national origin, religion, sex, pregnancy, marital status, sexual orientation, gender identity/expression, age, disability, genetic information, military service, covered/protected veteran status or any other federal, state or local protected class. At GSK, the health and safety of our employees are of paramount importance. As a science-led healthcare company on a mission to get ahead of disease together, we believe that supporting vaccination against COVID-19 is the single best thing we can do in the US to ensure the health and safety of our employees, complementary workers, workplaces, customers, consumers, communities, and the patients we serve. GSK has made the decision to require all US employees to be fully vaccinated against COVID-19, where allowed by state or local law and where vaccine supply is readily available. The only exceptions to this requirement are employees who are approved for an accommodation for religious, medical or disability-related reasons. Important notice to Employment businesses/ Agencies GSK does not accept referrals from employment businesses and/or employment agencies in respect of the vacancies posted on this site. All employment businesses/agencies are required to contact GSK's commercial and general procurement/human resources department to obtain prior written authorization before referring any candidates to GSK. The obtaining of prior written authorization is a condition precedent to any agreement (verbal or written) between the employment business/ agency and GSK. In the absence of such written authorization being obtained any actions undertaken by the employment business/agency shall be deemed to have been performed without the consent or contractual agreement of GSK. GSK shall therefore not be liable for any fees arising from such actions or any fees arising from any referrals by employment businesses/agencies in respect of the vacancies posted on this site. Please note that if you are a US Licensed Healthcare Professional or Healthcare Professional as defined by the laws of the state issuing your license, GSK may be required to capture and report expenses GSK incurs, on your behalf, in the event you are afforded an interview for employment. This capture of applicable transfers of value is necessary to ensure GSK's compliance to all federal and state US Transparency requirements. For more information, please visit GSK's Transparency Reporting For the Record site.
Location Whilst you may have any of our UK offices as a base location, you must be fully flexible in terms of assignment location, as these roles may involve periods of time away from home during the week at short notice. Capgemini requires our employees to be geographically mobile and to be able to travel to customer site to perform our jobs. Who you'll be working with The Cloud Data Platforms team is part of the Insights and Data Global Practice and has seen strong growth and continued success across a variety of projects and sectors. Cloud Data Platforms is the home of the Data Engineers, Platform Engineers, Solutions Architects and Business Analysts who are focused on driving our customers digital and data transformation journey using the modern cloud platforms. We specialise on using the latest frameworks, reference architectures and technologies using AWS, Azure and GCP. We continue to grow and are looking for talented individuals who want to join our high performing team. If you would like to develop your career as part of a team of highly skilled professionals who are passionate about increasing the value of the data and analytics in organisations you have come to the right place. The focus of your role We are looking for strong GCP Data Engineers who are passionate about Cloud technology and who ideally have skills in many of the following areas: • Build and deliver GCP data engineering solutions as part of a larger project • Use Google Data Products tools (e.g. BigQuery, Dataflow, Dataproc, AI Building Blocks, Looker, Cloud Data Fusion, Dataprep, etc.) to build solutions for our customers • Experience in Spark (Scala/Python/Java) and Kafka. • Experience in MDM, Metadata Management, Data Quality and Data Lineage tools. • E2E Data Engineering and Lifecycle (including non-functional requirements and operations) management. • E2E Solution Design skills - Prototyping, Usability testing and data visualization literacy. • Experience with SQL and NoSQL modern data stores. • Build relationships with client stakeholders to establish a high-level of rapport and confidence • Work with clients, local teams and offshore resources to deliver modern data products • Work effectively on client sites, Capgemini offices and from home • Use GCP Data focused Reference Architecture • Design and build data service APIs • Analyze current business practices, processes and procedures and identify future opportunities for leveraging GCP services • Design solutions and support the planning and implementation of data platform services including sizing, configuration, and needs assessment • Implement effective metrics and monitoring processes Skills Needed • Minimum 3-4 years of experience with Google Data Products tools (e.g. BigQuery, Dataflow, Dataproc, AI Building Blocks, Looker, Cloud Data Fusion, Dataprep, etc.) • Google Cloud Platform • Java, Scala, Python, Spark, SQL • Experience of developing enterprise grade ETL/ELT data pipelines. • Deep understanding of data manipulation/wrangling techniques • Demonstrable knowledge of applying Data Engineering best practices (coding practices to DS, unit testing, version control, code review). • Big Data Eco-Systems, Cloudera/Hortonworks, AWS EMR, GCP DataProc or GCP Cloud Data Fusion. • NoSQL Databases. Dynamo DB/Neo4j/Elastic, Google Cloud Datastore. • Snowflake Data Warehouse/Platform • Streaming technologies and processing engines, Kinesis, Kafka, Pub/Sub and Spark Streaming. • Experience of working CI/CD technologies, Git, Jenkins, Spinnaker, GCP Cloud Build, Ansible etc • Experience and knowledge of application Containerisation, Docker, Kubernetes etc • Experience building and deploying solutions to Cloud (AWS, Google Cloud) including Cloud provisioning tools • Strong interpersonal skills with the ability to work with clients to establish requirements in non-technical language. • Ability to translate business requirements into plausible technical solutions for articulation to other development staff. • Good understanding of Lambda architecture patterns • Good understanding of Data Governance, including Master Data Management (MDM) and Data Quality tools and processes • Influencing and supporting project delivery through involvement in project/sprint planning and QA • Experience with Agile methodology • Experience on collaboration tools such as JIRA, Kanban Board, Confluence etc Nice to Haves: • Knowledge of other cloud platforms • AWS (e.g Athena, Redshift, Glue, EMR) • Relevant certifications • Python • Snowflake • Databricks What we'll offer you Professional development. Accelerated career progression. An environment that encourages entrepreneurial spirit. It's all on offer at Capgemini and although collaboration is at the core of the way we work, we also recognise individual needs with a flexible benefits package you can tailor to suit you. Why we're different At Capgemini, we help organisations across the world become more agile, more competitive and more successful. Smart, tailored, often-groundbreaking technical solutions to complex problems are the norm. But so, too, is a culture that's as collaborative as it is forward thinking. Working closely with each other, and with our clients, we get under the skin of businesses and to the heart of their goals. You will too. Capgemini is proud to represent nearly 130 nationalities and its cultural diversity. Our holistic definition of diversity extends beyond gender, gender identity, sexual orientation, disability, ethnicity, race, age and religion. Capgemini views diversity as everything that makes us who we are as an organization, including our social background, our experiences in life and work, our communication styles and even our personality. These dimensions contribute to the type of diversity we value the most: diversity of thought. About Capgemini Capgemini is a global leader in partnering with companies to transform and manage their business by harnessing the power of technology. The Group is guided everyday by its purpose of unleashing human energy through technology for an inclusive and sustainable future. It is a responsible and diverse organization of 270,000 team members in nearly 50 countries. With its strong 50 year heritage and deep industry expertise, Capgemini is trusted by its clients to address the entire breadth of their business needs, from strategy and design to operations, fuelled by the fast evolving and innovative world of cloud, data, AI, connectivity, software, digital engineering and platforms. The Group reported in 2020 global revenues of €16 billion. Discover more about what Capgemini can offer you. Visit: and
Sep 23, 2022
Full time
Location Whilst you may have any of our UK offices as a base location, you must be fully flexible in terms of assignment location, as these roles may involve periods of time away from home during the week at short notice. Capgemini requires our employees to be geographically mobile and to be able to travel to customer site to perform our jobs. Who you'll be working with The Cloud Data Platforms team is part of the Insights and Data Global Practice and has seen strong growth and continued success across a variety of projects and sectors. Cloud Data Platforms is the home of the Data Engineers, Platform Engineers, Solutions Architects and Business Analysts who are focused on driving our customers digital and data transformation journey using the modern cloud platforms. We specialise on using the latest frameworks, reference architectures and technologies using AWS, Azure and GCP. We continue to grow and are looking for talented individuals who want to join our high performing team. If you would like to develop your career as part of a team of highly skilled professionals who are passionate about increasing the value of the data and analytics in organisations you have come to the right place. The focus of your role We are looking for strong GCP Data Engineers who are passionate about Cloud technology and who ideally have skills in many of the following areas: • Build and deliver GCP data engineering solutions as part of a larger project • Use Google Data Products tools (e.g. BigQuery, Dataflow, Dataproc, AI Building Blocks, Looker, Cloud Data Fusion, Dataprep, etc.) to build solutions for our customers • Experience in Spark (Scala/Python/Java) and Kafka. • Experience in MDM, Metadata Management, Data Quality and Data Lineage tools. • E2E Data Engineering and Lifecycle (including non-functional requirements and operations) management. • E2E Solution Design skills - Prototyping, Usability testing and data visualization literacy. • Experience with SQL and NoSQL modern data stores. • Build relationships with client stakeholders to establish a high-level of rapport and confidence • Work with clients, local teams and offshore resources to deliver modern data products • Work effectively on client sites, Capgemini offices and from home • Use GCP Data focused Reference Architecture • Design and build data service APIs • Analyze current business practices, processes and procedures and identify future opportunities for leveraging GCP services • Design solutions and support the planning and implementation of data platform services including sizing, configuration, and needs assessment • Implement effective metrics and monitoring processes Skills Needed • Minimum 3-4 years of experience with Google Data Products tools (e.g. BigQuery, Dataflow, Dataproc, AI Building Blocks, Looker, Cloud Data Fusion, Dataprep, etc.) • Google Cloud Platform • Java, Scala, Python, Spark, SQL • Experience of developing enterprise grade ETL/ELT data pipelines. • Deep understanding of data manipulation/wrangling techniques • Demonstrable knowledge of applying Data Engineering best practices (coding practices to DS, unit testing, version control, code review). • Big Data Eco-Systems, Cloudera/Hortonworks, AWS EMR, GCP DataProc or GCP Cloud Data Fusion. • NoSQL Databases. Dynamo DB/Neo4j/Elastic, Google Cloud Datastore. • Snowflake Data Warehouse/Platform • Streaming technologies and processing engines, Kinesis, Kafka, Pub/Sub and Spark Streaming. • Experience of working CI/CD technologies, Git, Jenkins, Spinnaker, GCP Cloud Build, Ansible etc • Experience and knowledge of application Containerisation, Docker, Kubernetes etc • Experience building and deploying solutions to Cloud (AWS, Google Cloud) including Cloud provisioning tools • Strong interpersonal skills with the ability to work with clients to establish requirements in non-technical language. • Ability to translate business requirements into plausible technical solutions for articulation to other development staff. • Good understanding of Lambda architecture patterns • Good understanding of Data Governance, including Master Data Management (MDM) and Data Quality tools and processes • Influencing and supporting project delivery through involvement in project/sprint planning and QA • Experience with Agile methodology • Experience on collaboration tools such as JIRA, Kanban Board, Confluence etc Nice to Haves: • Knowledge of other cloud platforms • AWS (e.g Athena, Redshift, Glue, EMR) • Relevant certifications • Python • Snowflake • Databricks What we'll offer you Professional development. Accelerated career progression. An environment that encourages entrepreneurial spirit. It's all on offer at Capgemini and although collaboration is at the core of the way we work, we also recognise individual needs with a flexible benefits package you can tailor to suit you. Why we're different At Capgemini, we help organisations across the world become more agile, more competitive and more successful. Smart, tailored, often-groundbreaking technical solutions to complex problems are the norm. But so, too, is a culture that's as collaborative as it is forward thinking. Working closely with each other, and with our clients, we get under the skin of businesses and to the heart of their goals. You will too. Capgemini is proud to represent nearly 130 nationalities and its cultural diversity. Our holistic definition of diversity extends beyond gender, gender identity, sexual orientation, disability, ethnicity, race, age and religion. Capgemini views diversity as everything that makes us who we are as an organization, including our social background, our experiences in life and work, our communication styles and even our personality. These dimensions contribute to the type of diversity we value the most: diversity of thought. About Capgemini Capgemini is a global leader in partnering with companies to transform and manage their business by harnessing the power of technology. The Group is guided everyday by its purpose of unleashing human energy through technology for an inclusive and sustainable future. It is a responsible and diverse organization of 270,000 team members in nearly 50 countries. With its strong 50 year heritage and deep industry expertise, Capgemini is trusted by its clients to address the entire breadth of their business needs, from strategy and design to operations, fuelled by the fast evolving and innovative world of cloud, data, AI, connectivity, software, digital engineering and platforms. The Group reported in 2020 global revenues of €16 billion. Discover more about what Capgemini can offer you. Visit: and
Site Name: UK - Hertfordshire - Stevenage, USA - Connecticut - Hartford, USA - Delaware - Dover, USA - Maryland - Rockville, USA - Massachusetts - Cambridge, USA - Massachusetts - Waltham, USA - New Jersey - Trenton, USA - Pennsylvania - Upper Providence Posted Date: Jun 6 2022 The mission of the Data Science and Data Engineering (DSDE) organization within GSK Pharmaceuticals R&D is to get the right data, to the right people, at the right time. TheData Framework and Opsorganization ensures we can do this efficiently, reliably, transparently, and at scale through the creation of a leading-edge, cloud-native data services framework. We focus heavily on developer experience, on strong, semantic abstractions for the data ecosystem, on professional operations and aggressive automation, and on transparency of operations and cost. Achieving delivery of the right data to the right people at the right time needs design and implementation of data flows and data products which leverage internal and external data assets and tools to drive discovery and development is a key objective for the Data Science and Data Engineering (DSDE) team within GSK's Pharmaceutical R&D organization. There are five key drivers for this approach, which are closely aligned with GSK's corporate priorities of Innovation, Performance and Trust: Automation of end-to-end data flows :Faster and reliable ingestion of high throughput data in genetics, genomics and multi-omics, to extract value of investments in new technology (instrument to analysis-ready data in Enabling governance by design of external and internal data :with engineered practical solutions for controlled use and monitoring Innovative disease-specific and domain-expert specific data products : to enable computational scientists and their research unit collaborators to get faster to key insights leading to faster biopharmaceutical development cycles. Supporting e2ecode traceability and data provenance :Increasing assurance of data integrity through automation, integration. Improving engineering efficiency :Extensible, reusable, scalable,updateable,maintainable, virtualized traceable data and code would be driven by data engineering innovation and better resource utilization. We are looking for experienced Senior DevOps Engineers to join our growing Data Ops team. As a Senior Dev Ops Engineer is a highly technical individual contributor, building modern, cloud-native, DevOps-first systems for standardizing and templatizingbiomedical and scientificdata engineering, with demonstrable experience across the following areas: Deliver declarative components for common data ingestion, transformation and publishing techniques Define and implement data governance aligned to modern standards Establish scalable, automated processes for data engineering teams across GSK Thought leader and partner with wider DSDE data engineering teams to advise on implementation and best practices Cloud Infrastructure-as-Code Define Service and Flow orchestration Data as a configurable resource(including configuration-driven access to scientific data modelling tools) Observability (monitoring, alerting, logging, tracing, etc.) Enable quality engineering through KPIs and code coverage and quality checks Standardise GitOps/declarative software development lifecycle Audit as a service Senior DevOpsEngineers take full ownership of delivering high-performing, high-impactbiomedical and scientificdataopsproducts and services, froma description of apattern thatcustomer Data Engineers are trying touseall the way through tofinal delivery (and ongoing monitoring and operations)of a templated project and all associated automation. They arestandard-bearers for software engineering and quality coding practices within theteam andareexpected to mentor more junior engineers; they may even coordinate the work of more junior engineers on a large project.Theydevise useful metrics for ensuring their services are meeting customer demand and having animpact anditerate to deliver and improve on those metrics in an agile fashion. Successful Senior DevOpsEngineers are developing expertise with the types of data and types of tools that are leveraged in the biomedical and scientific data engineering space, andhas the following skills and experience(withsignificant depth in one or more of these areas): Demonstrable experience deploying robust modularised/container-based solutions to production (ideally GCP) and leveraging the Cloud NativeComputing Foundation (CNCF) ecosystem Significant depth in DevOps principles and tools (e.g.GitOps, Jenkins,CircleCI, Azure DevOps, etc.), and how to integrate these tools with other productivity tools (e.g. Jira, Slack, Microsoft Teams) to build a comprehensive workflow Programming in Python. Scala orGo Embedding agile software engineering (task/issue management, testing, documentation, software development lifecycle, source control, etc.) Leveraging major cloud providers, both via Kubernetesorvia vendor-specific services Authentication and Authorization flows and associated technologies (e.g.OAuth2 + JWT) Common distributed data tools (e.g.Spark, Hive) The DSDE team is built on the principles of ownership, accountability, continuous development, and collaboration. We hire for the long term, and we're motivated to make this a great place to work. Our leaders will be committed to your career and development from day one. Why you? Basic Qualifications: We are looking for professionals with these required skills to achieve our goals: Masters in Computer Science with a focus in Data Engineering, DataOps, DevOps, MLOps, Software Engineering, etc, plus 5 years job experience (or PhD plus 3 years job experience) Experience with DevOps tools and concepts (e.g. Jira, GitLabs / Jenkins / CircleCI / Azure DevOps /etc.)Excellent with common distributed data tools in a production setting (Spark, Kafka, etc) Experience with specialized data architecture (e.g. optimizing physical layout for access patterns, including bloom filters, optimizing against self-describing formats such as ORC or Parquet, etc.) Experience with search / indexing systems (e.g. Elasticsearch) Expertise with agile development in Python, Scala, Go, and/or C++ Experience building reusable components on top of the CNCF ecosystem including Kubernetes Metrics-first mindset Experience mentoring junior engineers into deep technical expertise Preferred Qualifications: If you have the following characteristics, it would be a plus: Experience with agile software development Experience with building and designing a DevOps-first way of working Experience with building reusable components on top of the CNCF ecosystem including Kubernetes (or similar ecosystem ) LI-GSK Why GSK? Our values and expectationsare at the heart of everything we do and form an important part of our culture. These include Patient focus, Transparency, Respect, Integrity along with Courage, Accountability, Development, and Teamwork. As GSK focuses on our values and expectations and a culture of innovation, performance, and trust, the successful candidate will demonstrate the following capabilities: Operating at pace and agile decision making - using evidence and applying judgement to balance pace, rigour and risk. Committed to delivering high-quality results, overcoming challenges, focusing on what matters, execution. Continuously looking for opportunities to learn, build skills and share learning. Sustaining energy and wellbeing Building strong relationships and collaboration, honest and open conversations. Budgeting and cost consciousness As a company driven by our values of Patient focus, Transparency, Respect and Integrity, we know inclusion and diversity are essential for us to be able to succeed. We want all our colleagues to thrive at GSK bringing their unique experiences, ensuring they feel good and to keep growing their careers. As a candidate for a role, we want you to feel the same way. As an Equal Opportunity Employer, we are open to all talent. In the US, we also adhere to Affirmative Action principles. This ensures that all qualified applicants will receive equal consideration for employment without regard to race/ethnicity, colour, national origin, religion, gender, pregnancy, marital status, sexual orientation, gender identity/expression, age, disability, genetic information, military service, covered/protected veteran status or any other federal, state or local protected class ( US only). We believe in an agile working culture for all our roles. If flexibility is important to you, we encourage you to explore with our hiring team what the opportunities are. Please don't hesitate to contact us if you'd like to discuss any adjustments to our process which might help you demonstrate your strengths and capabilities. You can either call us on , or send an email As you apply, we will ask you to share some personal information which is entirely voluntary..... click apply for full job details
Sep 23, 2022
Full time
Site Name: UK - Hertfordshire - Stevenage, USA - Connecticut - Hartford, USA - Delaware - Dover, USA - Maryland - Rockville, USA - Massachusetts - Cambridge, USA - Massachusetts - Waltham, USA - New Jersey - Trenton, USA - Pennsylvania - Upper Providence Posted Date: Jun 6 2022 The mission of the Data Science and Data Engineering (DSDE) organization within GSK Pharmaceuticals R&D is to get the right data, to the right people, at the right time. TheData Framework and Opsorganization ensures we can do this efficiently, reliably, transparently, and at scale through the creation of a leading-edge, cloud-native data services framework. We focus heavily on developer experience, on strong, semantic abstractions for the data ecosystem, on professional operations and aggressive automation, and on transparency of operations and cost. Achieving delivery of the right data to the right people at the right time needs design and implementation of data flows and data products which leverage internal and external data assets and tools to drive discovery and development is a key objective for the Data Science and Data Engineering (DSDE) team within GSK's Pharmaceutical R&D organization. There are five key drivers for this approach, which are closely aligned with GSK's corporate priorities of Innovation, Performance and Trust: Automation of end-to-end data flows :Faster and reliable ingestion of high throughput data in genetics, genomics and multi-omics, to extract value of investments in new technology (instrument to analysis-ready data in Enabling governance by design of external and internal data :with engineered practical solutions for controlled use and monitoring Innovative disease-specific and domain-expert specific data products : to enable computational scientists and their research unit collaborators to get faster to key insights leading to faster biopharmaceutical development cycles. Supporting e2ecode traceability and data provenance :Increasing assurance of data integrity through automation, integration. Improving engineering efficiency :Extensible, reusable, scalable,updateable,maintainable, virtualized traceable data and code would be driven by data engineering innovation and better resource utilization. We are looking for experienced Senior DevOps Engineers to join our growing Data Ops team. As a Senior Dev Ops Engineer is a highly technical individual contributor, building modern, cloud-native, DevOps-first systems for standardizing and templatizingbiomedical and scientificdata engineering, with demonstrable experience across the following areas: Deliver declarative components for common data ingestion, transformation and publishing techniques Define and implement data governance aligned to modern standards Establish scalable, automated processes for data engineering teams across GSK Thought leader and partner with wider DSDE data engineering teams to advise on implementation and best practices Cloud Infrastructure-as-Code Define Service and Flow orchestration Data as a configurable resource(including configuration-driven access to scientific data modelling tools) Observability (monitoring, alerting, logging, tracing, etc.) Enable quality engineering through KPIs and code coverage and quality checks Standardise GitOps/declarative software development lifecycle Audit as a service Senior DevOpsEngineers take full ownership of delivering high-performing, high-impactbiomedical and scientificdataopsproducts and services, froma description of apattern thatcustomer Data Engineers are trying touseall the way through tofinal delivery (and ongoing monitoring and operations)of a templated project and all associated automation. They arestandard-bearers for software engineering and quality coding practices within theteam andareexpected to mentor more junior engineers; they may even coordinate the work of more junior engineers on a large project.Theydevise useful metrics for ensuring their services are meeting customer demand and having animpact anditerate to deliver and improve on those metrics in an agile fashion. Successful Senior DevOpsEngineers are developing expertise with the types of data and types of tools that are leveraged in the biomedical and scientific data engineering space, andhas the following skills and experience(withsignificant depth in one or more of these areas): Demonstrable experience deploying robust modularised/container-based solutions to production (ideally GCP) and leveraging the Cloud NativeComputing Foundation (CNCF) ecosystem Significant depth in DevOps principles and tools (e.g.GitOps, Jenkins,CircleCI, Azure DevOps, etc.), and how to integrate these tools with other productivity tools (e.g. Jira, Slack, Microsoft Teams) to build a comprehensive workflow Programming in Python. Scala orGo Embedding agile software engineering (task/issue management, testing, documentation, software development lifecycle, source control, etc.) Leveraging major cloud providers, both via Kubernetesorvia vendor-specific services Authentication and Authorization flows and associated technologies (e.g.OAuth2 + JWT) Common distributed data tools (e.g.Spark, Hive) The DSDE team is built on the principles of ownership, accountability, continuous development, and collaboration. We hire for the long term, and we're motivated to make this a great place to work. Our leaders will be committed to your career and development from day one. Why you? Basic Qualifications: We are looking for professionals with these required skills to achieve our goals: Masters in Computer Science with a focus in Data Engineering, DataOps, DevOps, MLOps, Software Engineering, etc, plus 5 years job experience (or PhD plus 3 years job experience) Experience with DevOps tools and concepts (e.g. Jira, GitLabs / Jenkins / CircleCI / Azure DevOps /etc.)Excellent with common distributed data tools in a production setting (Spark, Kafka, etc) Experience with specialized data architecture (e.g. optimizing physical layout for access patterns, including bloom filters, optimizing against self-describing formats such as ORC or Parquet, etc.) Experience with search / indexing systems (e.g. Elasticsearch) Expertise with agile development in Python, Scala, Go, and/or C++ Experience building reusable components on top of the CNCF ecosystem including Kubernetes Metrics-first mindset Experience mentoring junior engineers into deep technical expertise Preferred Qualifications: If you have the following characteristics, it would be a plus: Experience with agile software development Experience with building and designing a DevOps-first way of working Experience with building reusable components on top of the CNCF ecosystem including Kubernetes (or similar ecosystem ) LI-GSK Why GSK? Our values and expectationsare at the heart of everything we do and form an important part of our culture. These include Patient focus, Transparency, Respect, Integrity along with Courage, Accountability, Development, and Teamwork. As GSK focuses on our values and expectations and a culture of innovation, performance, and trust, the successful candidate will demonstrate the following capabilities: Operating at pace and agile decision making - using evidence and applying judgement to balance pace, rigour and risk. Committed to delivering high-quality results, overcoming challenges, focusing on what matters, execution. Continuously looking for opportunities to learn, build skills and share learning. Sustaining energy and wellbeing Building strong relationships and collaboration, honest and open conversations. Budgeting and cost consciousness As a company driven by our values of Patient focus, Transparency, Respect and Integrity, we know inclusion and diversity are essential for us to be able to succeed. We want all our colleagues to thrive at GSK bringing their unique experiences, ensuring they feel good and to keep growing their careers. As a candidate for a role, we want you to feel the same way. As an Equal Opportunity Employer, we are open to all talent. In the US, we also adhere to Affirmative Action principles. This ensures that all qualified applicants will receive equal consideration for employment without regard to race/ethnicity, colour, national origin, religion, gender, pregnancy, marital status, sexual orientation, gender identity/expression, age, disability, genetic information, military service, covered/protected veteran status or any other federal, state or local protected class ( US only). We believe in an agile working culture for all our roles. If flexibility is important to you, we encourage you to explore with our hiring team what the opportunities are. Please don't hesitate to contact us if you'd like to discuss any adjustments to our process which might help you demonstrate your strengths and capabilities. You can either call us on , or send an email As you apply, we will ask you to share some personal information which is entirely voluntary..... click apply for full job details
Site Name: UK - Hertfordshire - Stevenage, USA - Connecticut - Hartford, USA - Delaware - Dover, USA - Maryland - Rockville, USA - Massachusetts - Waltham, USA - Pennsylvania - Upper Providence, Warren NJ Posted Date: Aug The mission of the Data Science and Data Engineering (DSDE) organization within GSK Pharmaceuticals R&D is to get the right data, to the right people, at the right time. TheData Framework and Opsorganization ensures we can do this efficiently, reliably, transparently, and at scale through the creation of a leading-edge, cloud-native data services framework. We focus heavily on developer experience, on strong, semantic abstractions for the data ecosystem, on professional operations and aggressive automation, and on transparency of operations and cost. Achieving delivery of the right data to the right people at the right time needs design and implementation of data flows and data products which leverage internal and external data assets and tools to drive discovery and development is a key objective for the Data Science and Data Engineering (DS D E) team within GSK's Pharmaceutical R&D organisation . There are five key drivers for this approach, which are closely aligned with GSK's corporate priorities of Innovation, Performance and Trust: Automation of end-to-end data flows: Faster and reliable ingestion of high throughput data in genetics, genomics and multi-omics, to extract value of investments in new technology (instrument to analysis-ready data in Enabling governance by design of external and internal data: with engineered practical solutions for controlled use and monitoring Innovative disease-specific and domain-expert specific data products : to enable computational scientists and their research unit collaborators to get faster to key insights leading to faster biopharmaceutical development cycles. Supporting e2 e code traceability and data provenance: Increasing assurance of data integrity through automation, integration Improving engineering efficiency: Extensible, reusable, scalable,updateable,maintainable, virtualized traceable data and code would b e driven by data engineering innovation and better resource utilization. We are looking for an experienced Sr. Data Ops Engineer to join our growing Data Ops team. As a Sr. Data Ops Engineer is a highly technical individual contributor, building modern, cloud-native, DevOps-first systems for standardizing and templatizingbiomedical and scientificdata engineering, with demonstrable experience across the following areas : Deliver declarative components for common data ingestion, transformation and publishing techniques Define and implement data governance aligned to modern standards Establish scalable, automated processes for data engineering team s across GSK Thought leader and partner with wider DSDE data engineering teams to advise on implementation and best practices Cloud Infrastructure-as-Code D efine Service and Flow orchestration Data as a configurable resource(including configuration-driven access to scientific data modelling tools) Ob servabilty (monitoring, alerting, logging, tracing, ...) Enable quality engineering through KPIs and c ode coverage and quality checks Standardise GitOps /declarative software development lifecycle Audit as a service Sr. DataOpsEngineerstake full ownership of delivering high-performing, high-impactbiomedical and scientificdataopsproducts and services, froma description of apattern thatcustomer Data Engineers are trying touseall the way through tofinal delivery (and ongoing monitoring and operations)of a templated project and all associated automation. They arestandard-bearers for software engineering and quality coding practices within theteam andareexpected to mentor more junior engineers; they may even coordinate the work of more junior engineers on a large project.Theydevise useful metrics for ensuring their services are meeting customer demand and having animpact anditerate to deliver and improve on those metrics in an agile fashion. A successfulSr.DataOpsEngineeris developing expertise with the types of data and types of tools that are leveraged in the biomedical and scientific data engineering space, andhas the following skills and experience(withsignificant depth in one or more of these areas): Demonstrable experience deploying robust modularised/ container based solutions to production (ideally GCP) and leveraging the Cloud NativeComputing Foundation (CNCF) ecosystem Significant depth in DevOps principles and tools ( e.g. GitOps , Jenkins, CircleCI , Azure DevOps, ...), and how to integrate these tools with other productivity tools (e.g. Jira, Slack, Microsoft Teams) to build a comprehensive workflow P rogramming in Python. Scala or Go Embedding agile s oftware engineering ( task/issue management, testing, documentation, software development lifecycle, source control, ) Leveraging major cloud providers, both via Kubernetes or via vendor-specific services Authentication and Authorization flows and associated technologies ( e.g. OAuth2 + JWT) Common distributed data tools ( e.g. Spark, Hive) The DSDE team is built on the principles of ownership, accountability, continuous development, and collaboration. We hire for the long term, and we're motivated to make this a great place to work. Our leaders will be committed to your career and development from day one. Why you? Basic Qualifications: Bachelors degree in Computer Science with a focus in Data Engineering, DataOps, DevOps, MLOps, Software Engineering, etc, plus 7 years job experience or Masters degree with 5 Years of experience (or PhD plus 3 years job experience) Deep experience with DevOps tools and concepts ( e.g. Jira, GitLabs / Jenkins / CircleCI / Azure DevOps / ...) Excellent with common distributed data tools in a production setting (Spark, Kafka, etc) Experience with specialized data architecture ( e.g. optimizing physical layout for access patterns, including bloom filters, optimizing against self-describing formats such as ORC or Parquet, etc) Experience with search / indexing systems ( e.g. Elasticsearch) Deep expertise with agile development in Python, Scala, Go, and/or C++ Experience building reusable components on top of the CNCF ecosystem including Kubernetes Metrics-first mindset Experience mentoring junior engineers into deep technical expertise Preferred Qualifications: If you have the following characteristics, it would be a plus: Experience with agile software development Experience building and designing a DevOps-first way of working Demonstrated experience building reusable components on top of the CNCF ecosystem including Kubernetes (or similar ecosystem ) LI-GSK Why GSK? Our values and expectations are at the heart of everything we do and form an important part of our culture. These include Patient focus, Transparency, Respect, Integrity along with Courage, Accountability, Development, and Teamwork. As GSK focuses on our values and expectations and a culture of innovation, performance, and trust, the successful candidate will demonstrate the following capabilities: Operating at pace and agile decision making - using evidence and applying judgement to balance pace, rigour and risk. Committed to delivering high-quality results, overcoming challenges, focusing on what matters, execution. Continuously looking for opportunities to learn, build skills and share learning. Sustaining energy and wellbeing Building strong relationships and collaboration, honest and open conversations. Budgeting and cost consciousness As a company driven by our values of Patient focus, Transparency, Respect and Integrity, we know inclusion and diversity are essential for us to be able to succeed. We want all our colleagues to thrive at GSK bringing their unique experiences, ensuring they feel good and to keep growing their careers. As a candidate for a role, we want you to feel the same way. As an Equal Opportunity Employer, we are open to all talent. In the US, we also adhere to Affirmative Action principles. This ensures that all qualified applicants will receive equal consideration for employment without regard to neurodiversity, race/ethnicity, colour, national origin, religion, gender, pregnancy, marital status, sexual orientation, gender identity/expression, age, disability, genetic information, military service, covered/protected veteran status or any other federal, state or local protected class ( US only). We believe in an agile working culture for all our roles. If flexibility is important to you, we encourage you to explore with our hiring team what the opportunities are. Please don't hesitate to contact us if you'd like to discuss any adjustments to our process which might help you demonstrate your strengths and capabilities. You can either call us on , or send an email As you apply, we will ask you to share some personal information which is entirely voluntary..... click apply for full job details
Sep 23, 2022
Full time
Site Name: UK - Hertfordshire - Stevenage, USA - Connecticut - Hartford, USA - Delaware - Dover, USA - Maryland - Rockville, USA - Massachusetts - Waltham, USA - Pennsylvania - Upper Providence, Warren NJ Posted Date: Aug The mission of the Data Science and Data Engineering (DSDE) organization within GSK Pharmaceuticals R&D is to get the right data, to the right people, at the right time. TheData Framework and Opsorganization ensures we can do this efficiently, reliably, transparently, and at scale through the creation of a leading-edge, cloud-native data services framework. We focus heavily on developer experience, on strong, semantic abstractions for the data ecosystem, on professional operations and aggressive automation, and on transparency of operations and cost. Achieving delivery of the right data to the right people at the right time needs design and implementation of data flows and data products which leverage internal and external data assets and tools to drive discovery and development is a key objective for the Data Science and Data Engineering (DS D E) team within GSK's Pharmaceutical R&D organisation . There are five key drivers for this approach, which are closely aligned with GSK's corporate priorities of Innovation, Performance and Trust: Automation of end-to-end data flows: Faster and reliable ingestion of high throughput data in genetics, genomics and multi-omics, to extract value of investments in new technology (instrument to analysis-ready data in Enabling governance by design of external and internal data: with engineered practical solutions for controlled use and monitoring Innovative disease-specific and domain-expert specific data products : to enable computational scientists and their research unit collaborators to get faster to key insights leading to faster biopharmaceutical development cycles. Supporting e2 e code traceability and data provenance: Increasing assurance of data integrity through automation, integration Improving engineering efficiency: Extensible, reusable, scalable,updateable,maintainable, virtualized traceable data and code would b e driven by data engineering innovation and better resource utilization. We are looking for an experienced Sr. Data Ops Engineer to join our growing Data Ops team. As a Sr. Data Ops Engineer is a highly technical individual contributor, building modern, cloud-native, DevOps-first systems for standardizing and templatizingbiomedical and scientificdata engineering, with demonstrable experience across the following areas : Deliver declarative components for common data ingestion, transformation and publishing techniques Define and implement data governance aligned to modern standards Establish scalable, automated processes for data engineering team s across GSK Thought leader and partner with wider DSDE data engineering teams to advise on implementation and best practices Cloud Infrastructure-as-Code D efine Service and Flow orchestration Data as a configurable resource(including configuration-driven access to scientific data modelling tools) Ob servabilty (monitoring, alerting, logging, tracing, ...) Enable quality engineering through KPIs and c ode coverage and quality checks Standardise GitOps /declarative software development lifecycle Audit as a service Sr. DataOpsEngineerstake full ownership of delivering high-performing, high-impactbiomedical and scientificdataopsproducts and services, froma description of apattern thatcustomer Data Engineers are trying touseall the way through tofinal delivery (and ongoing monitoring and operations)of a templated project and all associated automation. They arestandard-bearers for software engineering and quality coding practices within theteam andareexpected to mentor more junior engineers; they may even coordinate the work of more junior engineers on a large project.Theydevise useful metrics for ensuring their services are meeting customer demand and having animpact anditerate to deliver and improve on those metrics in an agile fashion. A successfulSr.DataOpsEngineeris developing expertise with the types of data and types of tools that are leveraged in the biomedical and scientific data engineering space, andhas the following skills and experience(withsignificant depth in one or more of these areas): Demonstrable experience deploying robust modularised/ container based solutions to production (ideally GCP) and leveraging the Cloud NativeComputing Foundation (CNCF) ecosystem Significant depth in DevOps principles and tools ( e.g. GitOps , Jenkins, CircleCI , Azure DevOps, ...), and how to integrate these tools with other productivity tools (e.g. Jira, Slack, Microsoft Teams) to build a comprehensive workflow P rogramming in Python. Scala or Go Embedding agile s oftware engineering ( task/issue management, testing, documentation, software development lifecycle, source control, ) Leveraging major cloud providers, both via Kubernetes or via vendor-specific services Authentication and Authorization flows and associated technologies ( e.g. OAuth2 + JWT) Common distributed data tools ( e.g. Spark, Hive) The DSDE team is built on the principles of ownership, accountability, continuous development, and collaboration. We hire for the long term, and we're motivated to make this a great place to work. Our leaders will be committed to your career and development from day one. Why you? Basic Qualifications: Bachelors degree in Computer Science with a focus in Data Engineering, DataOps, DevOps, MLOps, Software Engineering, etc, plus 7 years job experience or Masters degree with 5 Years of experience (or PhD plus 3 years job experience) Deep experience with DevOps tools and concepts ( e.g. Jira, GitLabs / Jenkins / CircleCI / Azure DevOps / ...) Excellent with common distributed data tools in a production setting (Spark, Kafka, etc) Experience with specialized data architecture ( e.g. optimizing physical layout for access patterns, including bloom filters, optimizing against self-describing formats such as ORC or Parquet, etc) Experience with search / indexing systems ( e.g. Elasticsearch) Deep expertise with agile development in Python, Scala, Go, and/or C++ Experience building reusable components on top of the CNCF ecosystem including Kubernetes Metrics-first mindset Experience mentoring junior engineers into deep technical expertise Preferred Qualifications: If you have the following characteristics, it would be a plus: Experience with agile software development Experience building and designing a DevOps-first way of working Demonstrated experience building reusable components on top of the CNCF ecosystem including Kubernetes (or similar ecosystem ) LI-GSK Why GSK? Our values and expectations are at the heart of everything we do and form an important part of our culture. These include Patient focus, Transparency, Respect, Integrity along with Courage, Accountability, Development, and Teamwork. As GSK focuses on our values and expectations and a culture of innovation, performance, and trust, the successful candidate will demonstrate the following capabilities: Operating at pace and agile decision making - using evidence and applying judgement to balance pace, rigour and risk. Committed to delivering high-quality results, overcoming challenges, focusing on what matters, execution. Continuously looking for opportunities to learn, build skills and share learning. Sustaining energy and wellbeing Building strong relationships and collaboration, honest and open conversations. Budgeting and cost consciousness As a company driven by our values of Patient focus, Transparency, Respect and Integrity, we know inclusion and diversity are essential for us to be able to succeed. We want all our colleagues to thrive at GSK bringing their unique experiences, ensuring they feel good and to keep growing their careers. As a candidate for a role, we want you to feel the same way. As an Equal Opportunity Employer, we are open to all talent. In the US, we also adhere to Affirmative Action principles. This ensures that all qualified applicants will receive equal consideration for employment without regard to neurodiversity, race/ethnicity, colour, national origin, religion, gender, pregnancy, marital status, sexual orientation, gender identity/expression, age, disability, genetic information, military service, covered/protected veteran status or any other federal, state or local protected class ( US only). We believe in an agile working culture for all our roles. If flexibility is important to you, we encourage you to explore with our hiring team what the opportunities are. Please don't hesitate to contact us if you'd like to discuss any adjustments to our process which might help you demonstrate your strengths and capabilities. You can either call us on , or send an email As you apply, we will ask you to share some personal information which is entirely voluntary..... click apply for full job details
Our mission is to help organisations have more genuine relationships with their customers by using data and machine learning to understand them on a human level and enable marketing personalisation at scale. We're a well-funded scale-up business, having raised £30m from several notable PE investors. We've had an initial focus in the telecommunications sector, landing some of the world's leading telecommunications brands including Verizon in the US and O2 in the UK. Our clients typically have tens of millions of customers and billions of interaction events each day. We have developed our platform to handle data at this enormous scale efficiently. Intent HQ is unique because we've focused on understanding customer interaction data. We do the hard-specialised work leaving our clients free to focus on integrating the insight into their business, with one Client seeing an additional £26m annual revenue as a result of using our products and services. Intent HQ is an exciting place to work, with a team of 75-people strong (and growing!), we have offices in London, Barcelona and New York and collectively speak over 15 languages! The Opportunity We are looking for multiple experienced Mid/ Senior Scala Engineers to join our friendly Engineering team, working alongside colleagues in the UK, Spain and USA. You will work within one of our squads, dealing with data ingestion and storage processes, customer ML analytics platform and insights and visualisation dashboards. It's an exciting time for us as we are growing our team and looking for individuals who are passionate about technology, sensitive about client needs and collaborate well in a team environment. What is in our stack? You will work with a variety of technologies including: Scala to develop most of the backend code, including REST services, data pipelines, making use of FS2 functional streams data processes, Cats Effect, IO monad, Circe, Doobie, Http4s and other Typelevel stack libraries. Pair programming, testing and continuous integration and deployment as our daily development practices. Kafka, SQS as distributed messages systems. Github platform as git repository, issue tracker, continuous integration and package registry. Cassandra, PostgreSQL, Elasticsearch, S3, HDFS as the main data stores. Docker, Nomad, Terraform, Consul and Vault in the operational stack. Spark for transforming multi-terabyte sets of data into valuable insights. What are we looking for? You must have some Scala or Java (8 and up) functional programming experience. You have microservices architecture and testing practices experience. You understand distributed application architecture. Experience with using CI/ CD tools. You feel comfortable in an Agile/ Lean environment. You can communicate well, and are willing to learn, listen and share knowledge with your teammates. While we'd love if you have experience in some of our tech stack, don't be afraid by it - apply even if you don't, anybody can learn a new technology! What will you do? You will face challenging problems and will need to find new solutions for them. You'll help us give back to the open-source community by contributing to the tools we use, and releasing new ones, for the benefit of everybody! Pair programming, peer review. With us, you'll have many opportunities to improve and learn - if you care about collaboration, fault tolerance, trust, confidence, support, and experimentation, Intent HQ is the right place for you! Our Culture Working for Intent HQ, you have the opportunity to work with cutting edge technology in an environment where you are encouraged to think and act outside traditional methods. We are a growing business with a flat structure meaning everyone is visible and able to make an impact, not just in their own role, but across the business on a daily basis. We promote a collaborative, innovative and sociable culture, oriented towards performance and competitive advantage. Our DNA is made up of flexibility to all staff with a family friendly approach and focus on personal development. We have high expectations and pride ourselves on our cultural standards. Diversity & Inclusion Intent HQ is an equal opportunities employer with an ethos of commitment to promoting and practicing diversity, equality and inclusion at work. At IHQ, different perspectives, ideas and experiences are valued and respected, with fair and equal opportunities provided for all. Benefits & Perks Flexible working hours 26 days annual leave + bank holidays Employee Assistance Programme Company Pension Scheme Free breakfast (when in the office) Friday night socials - virtual and in-person (Covid permitting)! Macbook Pro Income protection
Sep 22, 2022
Full time
Our mission is to help organisations have more genuine relationships with their customers by using data and machine learning to understand them on a human level and enable marketing personalisation at scale. We're a well-funded scale-up business, having raised £30m from several notable PE investors. We've had an initial focus in the telecommunications sector, landing some of the world's leading telecommunications brands including Verizon in the US and O2 in the UK. Our clients typically have tens of millions of customers and billions of interaction events each day. We have developed our platform to handle data at this enormous scale efficiently. Intent HQ is unique because we've focused on understanding customer interaction data. We do the hard-specialised work leaving our clients free to focus on integrating the insight into their business, with one Client seeing an additional £26m annual revenue as a result of using our products and services. Intent HQ is an exciting place to work, with a team of 75-people strong (and growing!), we have offices in London, Barcelona and New York and collectively speak over 15 languages! The Opportunity We are looking for multiple experienced Mid/ Senior Scala Engineers to join our friendly Engineering team, working alongside colleagues in the UK, Spain and USA. You will work within one of our squads, dealing with data ingestion and storage processes, customer ML analytics platform and insights and visualisation dashboards. It's an exciting time for us as we are growing our team and looking for individuals who are passionate about technology, sensitive about client needs and collaborate well in a team environment. What is in our stack? You will work with a variety of technologies including: Scala to develop most of the backend code, including REST services, data pipelines, making use of FS2 functional streams data processes, Cats Effect, IO monad, Circe, Doobie, Http4s and other Typelevel stack libraries. Pair programming, testing and continuous integration and deployment as our daily development practices. Kafka, SQS as distributed messages systems. Github platform as git repository, issue tracker, continuous integration and package registry. Cassandra, PostgreSQL, Elasticsearch, S3, HDFS as the main data stores. Docker, Nomad, Terraform, Consul and Vault in the operational stack. Spark for transforming multi-terabyte sets of data into valuable insights. What are we looking for? You must have some Scala or Java (8 and up) functional programming experience. You have microservices architecture and testing practices experience. You understand distributed application architecture. Experience with using CI/ CD tools. You feel comfortable in an Agile/ Lean environment. You can communicate well, and are willing to learn, listen and share knowledge with your teammates. While we'd love if you have experience in some of our tech stack, don't be afraid by it - apply even if you don't, anybody can learn a new technology! What will you do? You will face challenging problems and will need to find new solutions for them. You'll help us give back to the open-source community by contributing to the tools we use, and releasing new ones, for the benefit of everybody! Pair programming, peer review. With us, you'll have many opportunities to improve and learn - if you care about collaboration, fault tolerance, trust, confidence, support, and experimentation, Intent HQ is the right place for you! Our Culture Working for Intent HQ, you have the opportunity to work with cutting edge technology in an environment where you are encouraged to think and act outside traditional methods. We are a growing business with a flat structure meaning everyone is visible and able to make an impact, not just in their own role, but across the business on a daily basis. We promote a collaborative, innovative and sociable culture, oriented towards performance and competitive advantage. Our DNA is made up of flexibility to all staff with a family friendly approach and focus on personal development. We have high expectations and pride ourselves on our cultural standards. Diversity & Inclusion Intent HQ is an equal opportunities employer with an ethos of commitment to promoting and practicing diversity, equality and inclusion at work. At IHQ, different perspectives, ideas and experiences are valued and respected, with fair and equal opportunities provided for all. Benefits & Perks Flexible working hours 26 days annual leave + bank holidays Employee Assistance Programme Company Pension Scheme Free breakfast (when in the office) Friday night socials - virtual and in-person (Covid permitting)! Macbook Pro Income protection
At NTT DATA are exceptional together and we believe in growing by helping others grow, clients and employees. We are constantly evolving, by creating more opportunities within our organisation and looking to innovative projects that shape your future. We provide all with the support you need to achieve the career of your dreams and we will give you the tools to become the best version of yourself you'd like to see. We are stronger by supporting people be who they want to be, work how they work best and envision the future through their own uniqueness. We provide a safe environment in which all of us can be ourselves and reach our full potential. This is a place where all individuals can be their true selves. Our success comes from our people, regardless of ethnicity, cultural background, gender, nationality, sexual orientation, or anything else that can be used to differentiate people because we are exceptional together. Because by supporting each other be who they want to be, work how they work best we bring innovation and we build more opportunities for everyone and a better future for our business, our people and our society. We understand you may not tick all the boxes of the requirements lists but if you are interested in this or other opportunities within NTT DATA, do not think twice and apply! As part of our continuous growth, we are looking for a NRT / Data Streaming Tech Lead to join our Data & Analytics team in London. We are looking for a hands-on tech lead/Architect to design, deliver, test, and troubleshoot complex data streaming implementations/NRT analytics. As a technical lead, you will work in partnership with our clients to deliver value, solve problems, and make a difference. You will be empowered to achieve your goals, and you will be involved in a wide array challenges and experiences giving you exposure to a variety of technologies, ensuring you always have the opportunity to learn and grow. Those who wish to benefit from long term career opportunities, within a prosperous Technology Consultancy environment, please get in touch. Develop solutions for streaming data ingestion, transformations and extraction in line with the streaming strategy. Experience working with streaming platforms, preferably Kafka Solid software engineering background with Java, Python or Scala You have experience designing large-scale data pipelines that utilise any stream-processing frameworks: Spark-Streaming, Kafka Streams, Flink Nice to have: Experience building and optimising big data pipelines, architectures, and data sets based on batch libraries (Spark, Flink,....) Experience with CI/CD tools. Build processes supporting data transformation, data structures, metadata, dependency and workload management. Working within the regulatory and system requirements to ensure data quality, data security, data compliance needs are met. Where you'll be: Our employees' safety is priority, so currently you will be working from the comfort of your own home most of the time. In the future when we can spend time in some more time of our offices, in Oxford Circus, Bank or Epworth House. As a result, you will be able to enjoy London's vibrant dynamics! .... And another thing, you will be joining a close-knit team that is supportive and approachable. Which means that as a new joiner, you will always have someone available to offer help and guidance. How to apply: We appreciate that you may not have an up to date CV, so please just send what you have and let's organise a chat!
Sep 21, 2022
Full time
At NTT DATA are exceptional together and we believe in growing by helping others grow, clients and employees. We are constantly evolving, by creating more opportunities within our organisation and looking to innovative projects that shape your future. We provide all with the support you need to achieve the career of your dreams and we will give you the tools to become the best version of yourself you'd like to see. We are stronger by supporting people be who they want to be, work how they work best and envision the future through their own uniqueness. We provide a safe environment in which all of us can be ourselves and reach our full potential. This is a place where all individuals can be their true selves. Our success comes from our people, regardless of ethnicity, cultural background, gender, nationality, sexual orientation, or anything else that can be used to differentiate people because we are exceptional together. Because by supporting each other be who they want to be, work how they work best we bring innovation and we build more opportunities for everyone and a better future for our business, our people and our society. We understand you may not tick all the boxes of the requirements lists but if you are interested in this or other opportunities within NTT DATA, do not think twice and apply! As part of our continuous growth, we are looking for a NRT / Data Streaming Tech Lead to join our Data & Analytics team in London. We are looking for a hands-on tech lead/Architect to design, deliver, test, and troubleshoot complex data streaming implementations/NRT analytics. As a technical lead, you will work in partnership with our clients to deliver value, solve problems, and make a difference. You will be empowered to achieve your goals, and you will be involved in a wide array challenges and experiences giving you exposure to a variety of technologies, ensuring you always have the opportunity to learn and grow. Those who wish to benefit from long term career opportunities, within a prosperous Technology Consultancy environment, please get in touch. Develop solutions for streaming data ingestion, transformations and extraction in line with the streaming strategy. Experience working with streaming platforms, preferably Kafka Solid software engineering background with Java, Python or Scala You have experience designing large-scale data pipelines that utilise any stream-processing frameworks: Spark-Streaming, Kafka Streams, Flink Nice to have: Experience building and optimising big data pipelines, architectures, and data sets based on batch libraries (Spark, Flink,....) Experience with CI/CD tools. Build processes supporting data transformation, data structures, metadata, dependency and workload management. Working within the regulatory and system requirements to ensure data quality, data security, data compliance needs are met. Where you'll be: Our employees' safety is priority, so currently you will be working from the comfort of your own home most of the time. In the future when we can spend time in some more time of our offices, in Oxford Circus, Bank or Epworth House. As a result, you will be able to enjoy London's vibrant dynamics! .... And another thing, you will be joining a close-knit team that is supportive and approachable. Which means that as a new joiner, you will always have someone available to offer help and guidance. How to apply: We appreciate that you may not have an up to date CV, so please just send what you have and let's organise a chat!
Site Name: USA - Pennsylvania - Upper Providence, UK - Hertfordshire - Stevenage, UK - London - Brentford, USA - Pennsylvania - Philadelphia Posted Date: Oct The mission of the Data Science and Data Engineering (DSDE) organization within GSK Pharmaceuticals R&D is to get the right data, to the right people, at the right time. TheData Framework and Opsorganization ensures we can do this efficiently, reliably, transparently, and at scale through the creation of a leading-edge, cloud-native data services framework. We focus heavily on developer experience, on strong, semantic abstractions for the data ecosystem, on professional operations and aggressive automation, and on transparency of operations and cost. We are looking for a skilled Data Ops Engineer II to join our growing team. The Data Ops team acceleratesbiomedicaland scientificdata product development and ensures consistent, professional-grade operations for the Data Science and Engineering organization by building templated projects (code repository plus DevOps pipelines) for various Data Science/Data Engineering architecture patternsin the challenging biomedical data space.A Data Ops Engineer IIknows the metrics desired for their tools andservices anditerates to deliver and improve on those metrics in an agile fashion. A Data Ops Engineer II is a highly technical individual contributor, building modern, cloud-native systems for standardizing and templatizing data engineering, such as: Standardized physical storage and search / indexing systems Schema management (data + metadata + versioning + provenance + governance) API semantics and ontology management Standard API architectures Kafka + standard streaming semantics Standard components for publishing data to file-based, relational, and other sorts of data stores Metadata systems Tooling for QA / evaluation Audit as a Service Additional responsibilities also include: Given a well-specified data framework problem, implement end-to-end solutionsusing appropriate programming languages(e.g.Python,Scala, or Go), open-source tools (e.g.Spark,Elasticsearch, ...), and cloud vendor-provided tools (e.g.Amazon S3) Leverage tools provided by Tech (e.g.infrastructure as code, CloudOps,DevOps, logging / alerting, ...) in delivery ofsolutions Write proper documentation in code as well as in wikis/other documentationsystems Writefantastic code along with theproper unit, functional, and integration tests for code and services to ensurequality Stayup to datewith developments in theopen-sourcecommunity around data engineering, data science, and similartooling The DSDE team is built on the principles of ownership, accountability, continuous development, and collaboration. We hire for the long term, and we're motivated to make this a great place to work. Our leaders will be committed to your career and development from day one. Why you? Basic Qualifications: We are looking for professionals with these required skills to achieve our goals: Master's in Computer Science with a focus in Data Engineering, DataOps, DevOps, MLOps, Software Engineering and 2+ years of experience OR PhD in Computer Science Demonstrated experience with software engineering (testing, documentation, software development lifecycle, source control, ... Experience with DevOps tools and concepts (e.g. Jira, GitLabs / Jenkins / CircleCI / Azure DevOps / ...) Experience with common distributed data tools in a production setting (Spark, Kafka, etc) Experience with basics of search engines/indexing (e.g. Elasticsearch, Lucene) Demonstrated experience in writing Python, Scala, Go, and/or C++ Preferred Qualifications: If you have the following characteristics, it would be a plus: Comfort with specialized data architecture (e.g. optimizing physical layout for access patterns, including bloom filters, optimizing against self-describing formats such as ORC or Parquet, etc) Experience with the CNCF ecosystem / Kubernetes Comfort with search/indexing systems (e.g. Elasticsearch) Experience with schema tools/schema management (Avro, Protobuf) Why GSK? Our values and expectations are at the heart of everything we do and form an important part of our culture. These include Patient focus, Transparency, Respect, Integrity along with Courage, Accountability, Development, and Teamwork. As GSK focuses on our values and expectations and a culture of innovation, performance, and trust, the successful candidate will demonstrate the following capabilities: Operating at pace and agile decision making - using evidence and applying judgement to balance pace, rigour and risk. Committed to delivering high-quality results, overcoming challenges, focusing on what matters, execution. Continuously looking for opportunities to learn, build skills and share learning. Sustaining energy and wellbeing Building strong relationships and collaboration, honest and open conversations. Budgeting and cost consciousness LI-GSK If you require an accommodation or other assistance to apply for a job at GSK, please contact the GSK Service Centre at 1- (US Toll Free) or +1 (outside US). GSK is an Equal Opportunity Employer and, in the US, we adhere to Affirmative Action principles. This ensures that all qualified applicants will receive equal consideration for employment without regard to race, color, national origin, religion, sex, pregnancy, marital status, sexual orientation, gender identity/expression, age, disability, genetic information, military service, covered/protected veteran status or any other federal, state or local protected class. At GSK, the health and safety of our employees are of paramount importance. As a science-led healthcare company on a mission to get ahead of disease together, we believe that supporting vaccination against COVID-19 is the single best thing we can do in the US to ensure the health and safety of our employees, complementary workers, workplaces, customers, consumers, communities, and the patients we serve. GSK has made the decision to require all US employees to be fully vaccinated against COVID-19, where allowed by state or local law and where vaccine supply is readily available. The only exceptions to this requirement are employees who are approved for an accommodation for religious, medical or disability-related reasons. Important notice to Employment businesses/ Agencies GSK does not accept referrals from employment businesses and/or employment agencies in respect of the vacancies posted on this site. All employment businesses/agencies are required to contact GSK's commercial and general procurement/human resources department to obtain prior written authorization before referring any candidates to GSK. The obtaining of prior written authorization is a condition precedent to any agreement (verbal or written) between the employment business/ agency and GSK. In the absence of such written authorization being obtained any actions undertaken by the employment business/agency shall be deemed to have been performed without the consent or contractual agreement of GSK. GSK shall therefore not be liable for any fees arising from such actions or any fees arising from any referrals by employment businesses/agencies in respect of the vacancies posted on this site. Please note that if you are a US Licensed Healthcare Professional or Healthcare Professional as defined by the laws of the state issuing your license, GSK may be required to capture and report expenses GSK incurs, on your behalf, in the event you are afforded an interview for employment. This capture of applicable transfers of value is necessary to ensure GSK's compliance to all federal and state US Transparency requirements. For more information, please visit GSK's Transparency Reporting For the Record site.
Sep 21, 2022
Full time
Site Name: USA - Pennsylvania - Upper Providence, UK - Hertfordshire - Stevenage, UK - London - Brentford, USA - Pennsylvania - Philadelphia Posted Date: Oct The mission of the Data Science and Data Engineering (DSDE) organization within GSK Pharmaceuticals R&D is to get the right data, to the right people, at the right time. TheData Framework and Opsorganization ensures we can do this efficiently, reliably, transparently, and at scale through the creation of a leading-edge, cloud-native data services framework. We focus heavily on developer experience, on strong, semantic abstractions for the data ecosystem, on professional operations and aggressive automation, and on transparency of operations and cost. We are looking for a skilled Data Ops Engineer II to join our growing team. The Data Ops team acceleratesbiomedicaland scientificdata product development and ensures consistent, professional-grade operations for the Data Science and Engineering organization by building templated projects (code repository plus DevOps pipelines) for various Data Science/Data Engineering architecture patternsin the challenging biomedical data space.A Data Ops Engineer IIknows the metrics desired for their tools andservices anditerates to deliver and improve on those metrics in an agile fashion. A Data Ops Engineer II is a highly technical individual contributor, building modern, cloud-native systems for standardizing and templatizing data engineering, such as: Standardized physical storage and search / indexing systems Schema management (data + metadata + versioning + provenance + governance) API semantics and ontology management Standard API architectures Kafka + standard streaming semantics Standard components for publishing data to file-based, relational, and other sorts of data stores Metadata systems Tooling for QA / evaluation Audit as a Service Additional responsibilities also include: Given a well-specified data framework problem, implement end-to-end solutionsusing appropriate programming languages(e.g.Python,Scala, or Go), open-source tools (e.g.Spark,Elasticsearch, ...), and cloud vendor-provided tools (e.g.Amazon S3) Leverage tools provided by Tech (e.g.infrastructure as code, CloudOps,DevOps, logging / alerting, ...) in delivery ofsolutions Write proper documentation in code as well as in wikis/other documentationsystems Writefantastic code along with theproper unit, functional, and integration tests for code and services to ensurequality Stayup to datewith developments in theopen-sourcecommunity around data engineering, data science, and similartooling The DSDE team is built on the principles of ownership, accountability, continuous development, and collaboration. We hire for the long term, and we're motivated to make this a great place to work. Our leaders will be committed to your career and development from day one. Why you? Basic Qualifications: We are looking for professionals with these required skills to achieve our goals: Master's in Computer Science with a focus in Data Engineering, DataOps, DevOps, MLOps, Software Engineering and 2+ years of experience OR PhD in Computer Science Demonstrated experience with software engineering (testing, documentation, software development lifecycle, source control, ... Experience with DevOps tools and concepts (e.g. Jira, GitLabs / Jenkins / CircleCI / Azure DevOps / ...) Experience with common distributed data tools in a production setting (Spark, Kafka, etc) Experience with basics of search engines/indexing (e.g. Elasticsearch, Lucene) Demonstrated experience in writing Python, Scala, Go, and/or C++ Preferred Qualifications: If you have the following characteristics, it would be a plus: Comfort with specialized data architecture (e.g. optimizing physical layout for access patterns, including bloom filters, optimizing against self-describing formats such as ORC or Parquet, etc) Experience with the CNCF ecosystem / Kubernetes Comfort with search/indexing systems (e.g. Elasticsearch) Experience with schema tools/schema management (Avro, Protobuf) Why GSK? Our values and expectations are at the heart of everything we do and form an important part of our culture. These include Patient focus, Transparency, Respect, Integrity along with Courage, Accountability, Development, and Teamwork. As GSK focuses on our values and expectations and a culture of innovation, performance, and trust, the successful candidate will demonstrate the following capabilities: Operating at pace and agile decision making - using evidence and applying judgement to balance pace, rigour and risk. Committed to delivering high-quality results, overcoming challenges, focusing on what matters, execution. Continuously looking for opportunities to learn, build skills and share learning. Sustaining energy and wellbeing Building strong relationships and collaboration, honest and open conversations. Budgeting and cost consciousness LI-GSK If you require an accommodation or other assistance to apply for a job at GSK, please contact the GSK Service Centre at 1- (US Toll Free) or +1 (outside US). GSK is an Equal Opportunity Employer and, in the US, we adhere to Affirmative Action principles. This ensures that all qualified applicants will receive equal consideration for employment without regard to race, color, national origin, religion, sex, pregnancy, marital status, sexual orientation, gender identity/expression, age, disability, genetic information, military service, covered/protected veteran status or any other federal, state or local protected class. At GSK, the health and safety of our employees are of paramount importance. As a science-led healthcare company on a mission to get ahead of disease together, we believe that supporting vaccination against COVID-19 is the single best thing we can do in the US to ensure the health and safety of our employees, complementary workers, workplaces, customers, consumers, communities, and the patients we serve. GSK has made the decision to require all US employees to be fully vaccinated against COVID-19, where allowed by state or local law and where vaccine supply is readily available. The only exceptions to this requirement are employees who are approved for an accommodation for religious, medical or disability-related reasons. Important notice to Employment businesses/ Agencies GSK does not accept referrals from employment businesses and/or employment agencies in respect of the vacancies posted on this site. All employment businesses/agencies are required to contact GSK's commercial and general procurement/human resources department to obtain prior written authorization before referring any candidates to GSK. The obtaining of prior written authorization is a condition precedent to any agreement (verbal or written) between the employment business/ agency and GSK. In the absence of such written authorization being obtained any actions undertaken by the employment business/agency shall be deemed to have been performed without the consent or contractual agreement of GSK. GSK shall therefore not be liable for any fees arising from such actions or any fees arising from any referrals by employment businesses/agencies in respect of the vacancies posted on this site. Please note that if you are a US Licensed Healthcare Professional or Healthcare Professional as defined by the laws of the state issuing your license, GSK may be required to capture and report expenses GSK incurs, on your behalf, in the event you are afforded an interview for employment. This capture of applicable transfers of value is necessary to ensure GSK's compliance to all federal and state US Transparency requirements. For more information, please visit GSK's Transparency Reporting For the Record site.
Scala/Python Development Lead: London - Remote (1-2 days in the office) Our client a global leader in providing financial service data and insight to assist some of the worlds leading organisation to make better financial decisions, is looking for a Scala/Python Development Lead . As the Scala/Python Development Lead you will be leading a diverse and global team of quantitative analysts and software developers. The level of the Scala/Python Development Lead is flexible, but Python/Scala, Linux/Unix, Git, AWS (ECS2, EMR, S3), Hadoop, NoSQL Databases, Kafka , Druid, SQL and SOLR experience would be required. As the Scala/Python Development Lead you will build state of the art financial analytics software using a Hadoop/Spark platform. The Scala/Python Development Lead will be working 50% on Development and 50% on leading the team/projects. As the Scala/Python Development Lead you will lead teams in design and development of Big Data infrastructure for batch and Real Time analytics. You as the Scala/Python Development Lead will need to be able to interpret and analyse business use-cases and feature requests into technical designs and development tasks. As the Scala/Python Development Lead you will need to ensure highly interactive response times to avoid performance bottlenecks and be active in system architecture and design discussions. Requirements: Min of 5-8 years of experience as a Scala/Python Development Lead. (Either Scala and Python) Work experience in the following: Python OR Scala AWS (ECS2, EMR,S3) Linux OR Unix GIT SQL Hadoop, HDFS, YARN, Spark, Hive, NoSQL databases, SOLR, Kafka, Druid Akka Note. The level is slightly flexible so if you're slightly less or more experienced as a Scala/Python Development Lead please do reach out. Salary - up to £125,000 + 10% Bonus + Benefits. Location - London - Remote mostly (1-2 days in the office) Duration - Permanent. Reply to this email with your CV attached and I shall call in the next 24 hours to discuss your skills and suitability in more detail. We're also supporting this client with similar roles such as DevOps Engineer with the same client, so please feel free to share this with your network.
Sep 20, 2022
Full time
Scala/Python Development Lead: London - Remote (1-2 days in the office) Our client a global leader in providing financial service data and insight to assist some of the worlds leading organisation to make better financial decisions, is looking for a Scala/Python Development Lead . As the Scala/Python Development Lead you will be leading a diverse and global team of quantitative analysts and software developers. The level of the Scala/Python Development Lead is flexible, but Python/Scala, Linux/Unix, Git, AWS (ECS2, EMR, S3), Hadoop, NoSQL Databases, Kafka , Druid, SQL and SOLR experience would be required. As the Scala/Python Development Lead you will build state of the art financial analytics software using a Hadoop/Spark platform. The Scala/Python Development Lead will be working 50% on Development and 50% on leading the team/projects. As the Scala/Python Development Lead you will lead teams in design and development of Big Data infrastructure for batch and Real Time analytics. You as the Scala/Python Development Lead will need to be able to interpret and analyse business use-cases and feature requests into technical designs and development tasks. As the Scala/Python Development Lead you will need to ensure highly interactive response times to avoid performance bottlenecks and be active in system architecture and design discussions. Requirements: Min of 5-8 years of experience as a Scala/Python Development Lead. (Either Scala and Python) Work experience in the following: Python OR Scala AWS (ECS2, EMR,S3) Linux OR Unix GIT SQL Hadoop, HDFS, YARN, Spark, Hive, NoSQL databases, SOLR, Kafka, Druid Akka Note. The level is slightly flexible so if you're slightly less or more experienced as a Scala/Python Development Lead please do reach out. Salary - up to £125,000 + 10% Bonus + Benefits. Location - London - Remote mostly (1-2 days in the office) Duration - Permanent. Reply to this email with your CV attached and I shall call in the next 24 hours to discuss your skills and suitability in more detail. We're also supporting this client with similar roles such as DevOps Engineer with the same client, so please feel free to share this with your network.
Data Engineer London, Edinburgh or Manchester - Remote Initially What is a Data Engineer? Dufrain consider a data engineer to be a multi-skilled individual with experience in delivering solutions across the technical data landscape. Our data engineers have expertise in technical delivery, technologies and concepts in areas such as Data Storage, Data Ingestion, Data Integration, Data Warehousing, Data Preparation and Cloud Infrastructure. A suitable candidate will also have a wider understanding of how their role and delivery contributes to wider business outcomes and be able to concisely articulate to stakeholders and interested parties their role and solutions in a way that can be easily understood. Essential Requirements: • 2+ years of experience in a data engineer role • Expert SQL Skills • Experience working with the Hadoop ecosystem (Spark, Hive/Impala, Databricks) • Strong Programming background in Java, Scala, R or Python • Experience in designing, developing and managing data pipelines to process large amounts of dataExperience of cloud based big data offerings such as Amazon EMR, Azure HDInsights or Google Dataproc • Experience of cloud platforms such as Azure, AWS or GCP • Experience working with an ETL tool e.g. Azure Data Factory, SSIS & Talend • Understanding of data modelling techniques e.g. Kimball • Experience working with data streams e.g. Kafka • Experience working with large, structured and unstructured datasets • Experience with database technologies (relational, NoSQL, graph) • Expertise with version control e.g. git • Flexibility to travel to and work on client sites within the UK and occasionally Europe. • Excellent track record in executive stakeholder and sponsor management and maintaining valuable relationships. Key Skills: • Takes ownership and accountability for mission critical initiatives and deliverables both internally and for clients • A proven leader of people having significant influence on the careers and aspirations of those working in proximity • Awareness of current market trends in data having the ability to influence opinion and decisioning across the Data Management spectrum #LifeAtDufrain #GetBusyLiving We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.
Feb 04, 2022
Full time
Data Engineer London, Edinburgh or Manchester - Remote Initially What is a Data Engineer? Dufrain consider a data engineer to be a multi-skilled individual with experience in delivering solutions across the technical data landscape. Our data engineers have expertise in technical delivery, technologies and concepts in areas such as Data Storage, Data Ingestion, Data Integration, Data Warehousing, Data Preparation and Cloud Infrastructure. A suitable candidate will also have a wider understanding of how their role and delivery contributes to wider business outcomes and be able to concisely articulate to stakeholders and interested parties their role and solutions in a way that can be easily understood. Essential Requirements: • 2+ years of experience in a data engineer role • Expert SQL Skills • Experience working with the Hadoop ecosystem (Spark, Hive/Impala, Databricks) • Strong Programming background in Java, Scala, R or Python • Experience in designing, developing and managing data pipelines to process large amounts of dataExperience of cloud based big data offerings such as Amazon EMR, Azure HDInsights or Google Dataproc • Experience of cloud platforms such as Azure, AWS or GCP • Experience working with an ETL tool e.g. Azure Data Factory, SSIS & Talend • Understanding of data modelling techniques e.g. Kimball • Experience working with data streams e.g. Kafka • Experience working with large, structured and unstructured datasets • Experience with database technologies (relational, NoSQL, graph) • Expertise with version control e.g. git • Flexibility to travel to and work on client sites within the UK and occasionally Europe. • Excellent track record in executive stakeholder and sponsor management and maintaining valuable relationships. Key Skills: • Takes ownership and accountability for mission critical initiatives and deliverables both internally and for clients • A proven leader of people having significant influence on the careers and aspirations of those working in proximity • Awareness of current market trends in data having the ability to influence opinion and decisioning across the Data Management spectrum #LifeAtDufrain #GetBusyLiving We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.
Data Engineer with PySpark, Kafka, Python and SQL experience is required for a leading financial services organisation based in London, to work on a greenfield project. This role is fully remote and is inside IR35 Strong commercial Data Engineering skills including Hadoop, Spark, Primary programming language is Java, Scala or Python Hands-on experience of Spark Development(preferably python). Expert knowledge of PySpark SQL Kafka Streaming Experience with python coding and packaging. Expert Hands-on with Continuous Integration, DevOps, Jenkins, TeamCity, Bit-Bucket/GIT, Artifactory Knowledge of working in AWS environment/other public cloud platforms Must have experience in design, implementation and DevOps. Demonstrable experience working on and delivering large scale development projects Strong background working on complex distributed system
Nov 05, 2021
Contractor
Data Engineer with PySpark, Kafka, Python and SQL experience is required for a leading financial services organisation based in London, to work on a greenfield project. This role is fully remote and is inside IR35 Strong commercial Data Engineering skills including Hadoop, Spark, Primary programming language is Java, Scala or Python Hands-on experience of Spark Development(preferably python). Expert knowledge of PySpark SQL Kafka Streaming Experience with python coding and packaging. Expert Hands-on with Continuous Integration, DevOps, Jenkins, TeamCity, Bit-Bucket/GIT, Artifactory Knowledge of working in AWS environment/other public cloud platforms Must have experience in design, implementation and DevOps. Demonstrable experience working on and delivering large scale development projects Strong background working on complex distributed system
About ExpressVPN We're a fast-growing SAAS tech company providing privacy and security protections to consumers around the world. We build and operate services that are easy to use, fast, and secure. We've been in business for more than 11 years, we're highly profitable, growing, and actively hiring . Some highlights: Large customer base worldwide , which brings many challenges and opportunities. We have millions of users who rely on our service 24/7/365 for a broad range of use-cases. Excellence in operations is therefore core to our business . Successful business in a fast-growing industry. We are a leader in a fragmented market and are growing quickly both because the industry is expanding and because we are gaining market share. We have sufficient financial resources to tackle large challenges for the long term. Highly skilled team using efficient processes. Meritocratic culture with fast decision-making. About the Marketing Attribution and Customer Engagement Tribe We are accountable for delivering: Web products and services to acquire, retain, and regain customers . Tools and services for customer relationship management , enabling marketers to communicate with customers across all apps and media. Tools and services for content management for our website and other media. Sales attribution tools and services. Services for customer support agents . We provide them with tools for interacting with our customer database (aka our "Backend UI") Split testing services to allow continuous improvements across all our communications and marketing activities. They rely on us for appropriately anonymized data. Mission-critical data processing that enables the Marketing tribe to increase their budgets with confidence and make sure our partners are reliably rewarded for their efforts. Our responsibilities include: Working with other data teams to build and operate data platform services, including defining and tracking its SLA. Guiding various engineering teams to design models and schemas of the data to be fed into the platform, making sure they can be processed in a scalable way and used by analysts efficiently. Guiding data analysts on the use of the data platform. Building libraries/modules and reference implementations of data ingesters on several common tech stacks. Guarding user privacy . While all teams are responsible for ensuring compliance of their work with our privacy policy, our team also has a veto right against processing any data that might not be compliant. Partnering with other teams on projects to build data engineering solutions such as for churn-prediction, payment fraud management, and other company-wide challenges. Other notes about our team: Our tech stack currently is mostly AWS Redshift, BigQuery, Spark, Airflow and Tableau, and we imagine it will evolve significantly over time. We intend to have a range of engineering roles on the team, covering people with backgrounds in software development, infrastructure operations, and data science. Responsibilities of this position This is a newly created position to become the leader of this team. The company's data pipelines and responsible engineers are currently spread across multiple teams. You joining the company will mark the beginning of a dedicated data engineering team to support processing the enormous amount of data we collect through all of our marketing properties. Your responsibilities will include: Create the hiring plan for your team world-wide, design plans for assessing candidates, work with our recruiting-team to source and test candidates, hire and onboard new staff. We're currently very much under-staffed. You need to grow the team significantly. Be the people leader for all members of the team . You hire, train, motivate, assign responsibilities, and hold people accountable. You will compare actual versus expected deliverables at high frequencies, and exchange actionable feedback. Articulate a strategic vision beyond just Marketing data, helping define best practices for cataloging and processing data at scale and making it available to internal and external stakeholders. Design processes , document them very clearly, measuring their metrics, and improving them. Be able to perform many of the duties of your team members. Be able to lead by example and be a credible advisor for all roles on your team. Desired skills and experience Be excellent at Data Engineering . Extensive experience operating large scale data pipelines and databases Extensive experience designing and implementing complicated ETL jobs Expert in SQL Experience in stream processing using tools like AWS Kinesis or Kafka Experience in big data analytic frameworks like Hadoop or Spark (optional) Experience in writing automated tests Enjoy managing people and be very good at it. That means you've had several direct reports for at least two years. you've demonstrated the ability to develop a team including hiring, coaching and mentoring of team members Be an experienced agile project manager . Requirements gathering and solution architecture . That means you've worked with multiple stakeholders to draw out requirements and define use cases you have a proven ability to assess uses cases and requirements, then design suitable solution architectures Strong English written and verbal communication skills. What we offer Challenging work in a fun and collaborative environment Attractive compensation and time-off benefits Full-time employment with flexible working hours Team lunches and company events every quarter Note: Please upload your resume as a PDF and do not include any salary or compensation information in it.
Oct 07, 2021
Full time
About ExpressVPN We're a fast-growing SAAS tech company providing privacy and security protections to consumers around the world. We build and operate services that are easy to use, fast, and secure. We've been in business for more than 11 years, we're highly profitable, growing, and actively hiring . Some highlights: Large customer base worldwide , which brings many challenges and opportunities. We have millions of users who rely on our service 24/7/365 for a broad range of use-cases. Excellence in operations is therefore core to our business . Successful business in a fast-growing industry. We are a leader in a fragmented market and are growing quickly both because the industry is expanding and because we are gaining market share. We have sufficient financial resources to tackle large challenges for the long term. Highly skilled team using efficient processes. Meritocratic culture with fast decision-making. About the Marketing Attribution and Customer Engagement Tribe We are accountable for delivering: Web products and services to acquire, retain, and regain customers . Tools and services for customer relationship management , enabling marketers to communicate with customers across all apps and media. Tools and services for content management for our website and other media. Sales attribution tools and services. Services for customer support agents . We provide them with tools for interacting with our customer database (aka our "Backend UI") Split testing services to allow continuous improvements across all our communications and marketing activities. They rely on us for appropriately anonymized data. Mission-critical data processing that enables the Marketing tribe to increase their budgets with confidence and make sure our partners are reliably rewarded for their efforts. Our responsibilities include: Working with other data teams to build and operate data platform services, including defining and tracking its SLA. Guiding various engineering teams to design models and schemas of the data to be fed into the platform, making sure they can be processed in a scalable way and used by analysts efficiently. Guiding data analysts on the use of the data platform. Building libraries/modules and reference implementations of data ingesters on several common tech stacks. Guarding user privacy . While all teams are responsible for ensuring compliance of their work with our privacy policy, our team also has a veto right against processing any data that might not be compliant. Partnering with other teams on projects to build data engineering solutions such as for churn-prediction, payment fraud management, and other company-wide challenges. Other notes about our team: Our tech stack currently is mostly AWS Redshift, BigQuery, Spark, Airflow and Tableau, and we imagine it will evolve significantly over time. We intend to have a range of engineering roles on the team, covering people with backgrounds in software development, infrastructure operations, and data science. Responsibilities of this position This is a newly created position to become the leader of this team. The company's data pipelines and responsible engineers are currently spread across multiple teams. You joining the company will mark the beginning of a dedicated data engineering team to support processing the enormous amount of data we collect through all of our marketing properties. Your responsibilities will include: Create the hiring plan for your team world-wide, design plans for assessing candidates, work with our recruiting-team to source and test candidates, hire and onboard new staff. We're currently very much under-staffed. You need to grow the team significantly. Be the people leader for all members of the team . You hire, train, motivate, assign responsibilities, and hold people accountable. You will compare actual versus expected deliverables at high frequencies, and exchange actionable feedback. Articulate a strategic vision beyond just Marketing data, helping define best practices for cataloging and processing data at scale and making it available to internal and external stakeholders. Design processes , document them very clearly, measuring their metrics, and improving them. Be able to perform many of the duties of your team members. Be able to lead by example and be a credible advisor for all roles on your team. Desired skills and experience Be excellent at Data Engineering . Extensive experience operating large scale data pipelines and databases Extensive experience designing and implementing complicated ETL jobs Expert in SQL Experience in stream processing using tools like AWS Kinesis or Kafka Experience in big data analytic frameworks like Hadoop or Spark (optional) Experience in writing automated tests Enjoy managing people and be very good at it. That means you've had several direct reports for at least two years. you've demonstrated the ability to develop a team including hiring, coaching and mentoring of team members Be an experienced agile project manager . Requirements gathering and solution architecture . That means you've worked with multiple stakeholders to draw out requirements and define use cases you have a proven ability to assess uses cases and requirements, then design suitable solution architectures Strong English written and verbal communication skills. What we offer Challenging work in a fun and collaborative environment Attractive compensation and time-off benefits Full-time employment with flexible working hours Team lunches and company events every quarter Note: Please upload your resume as a PDF and do not include any salary or compensation information in it.
Ampersand Consulting
Ampersand Consulting
Senior Data Engineer (Cloud, Azure, SQL, Python, Java) The UK's leading provider of IT solutions and services is looking to hire an experienced Senior Data Engineer (Cloud, Azure, SQL, Python, Java) to join their rapidly expanding Data Analytics team working closely with the Chief Data Scientist based in Central London. The Senior Data Engineer (Cloud, Azure, SQL, Python, Java) will implement complex multi/hybrid cloud based big data projects with a focus on collecting, parsing, managing, analyzing and visualizing large sets of data to turn information into insights using multiple technology platforms. The Senior Data Engineer (Cloud, Azure, SQL, Python, Java) will be a technical contributor with hands-on knowledge of all phases in building large-scale cloud based distributed data processing systems and applications. Responsibilities for the Senior Data Engineer (Cloud, Azure, SQL, Python, Java) Lead the design, implementation, and continuous delivery of pipelines using distributed Azure based big data technologies supporting data processing initiatives across batch and streaming datasets Experience in hybrid cloud solutions and experience in GCP resources Responsible for development using Scala, Python languages and Big Data Frameworks such as Spark, EMR, Kafka, Storm, Jenkins, Jfrog Artifactory and DataBricks Provide administrative support on deployed Azure platform components Identify, evaluate and implement cutting edge big data pipelines and frameworks required to provide requested capabilities to integrate external data sources and APIs Review, analyse and evaluate market requirements, business requirements and project briefs in order to design the most appropriate end-to-end technology solutions Process and manage high volume real time customer interaction streams Provide architectural support by building Proof of Concepts & Prototypes Self-Starter to deliver data engineering solutions to optimize both the cost and existing solution Requirements for the Senior Data Engineer (Cloud, Azure, SQL, Python, Java) Extensive Software Industry experience Development experience with Azure services (such as Data Factories, Data Lake Storage, SQL Elastic Pools, Data Pipeline, DataBricks, Kubernetes Services, Apache Nifi, jFrog etc) Development experience with GCP services (Cloud Storage, Cloud Spanner, BigQuery) Experience with Apache Spark and NoSQL Implementation Extensive working knowledge in different programming or scripting languages like Scala, Java, Linux, Shell, SQL, Python Proficiency working with structured, semi-structured and unstructured data sets including social, web logs and real time streaming data feeds Able to tune Big Data solutions to improve performance and end-user experience Knowledge on Visualization and Data Science Tools Expert level usage with Jenkins, GitHub is preferred If you would like to apply for the Senior Data Engineer (Cloud, Azure, SQL, Python, Java) role then please click the apply button
Feb 15, 2019
Ampersand Consulting
Ampersand Consulting
Senior Data Engineer (Cloud, Azure, SQL, Python, Java) The UK's leading provider of IT solutions and services is looking to hire an experienced Senior Data Engineer (Cloud, Azure, SQL, Python, Java) to join their rapidly expanding Data Analytics team working closely with the Chief Data Scientist based in Central London. The Senior Data Engineer (Cloud, Azure, SQL, Python, Java) will implement complex multi/hybrid cloud based big data projects with a focus on collecting, parsing, managing, analyzing and visualizing large sets of data to turn information into insights using multiple technology platforms. The Senior Data Engineer (Cloud, Azure, SQL, Python, Java) will be a technical contributor with hands-on knowledge of all phases in building large-scale cloud based distributed data processing systems and applications. Responsibilities for the Senior Data Engineer (Cloud, Azure, SQL, Python, Java) Lead the design, implementation, and continuous delivery of pipelines using distributed Azure based big data technologies supporting data processing initiatives across batch and streaming datasets Experience in hybrid cloud solutions and experience in GCP resources Responsible for development using Scala, Python languages and Big Data Frameworks such as Spark, EMR, Kafka, Storm, Jenkins, Jfrog Artifactory and DataBricks Provide administrative support on deployed Azure platform components Identify, evaluate and implement cutting edge big data pipelines and frameworks required to provide requested capabilities to integrate external data sources and APIs Review, analyse and evaluate market requirements, business requirements and project briefs in order to design the most appropriate end-to-end technology solutions Process and manage high volume real time customer interaction streams Provide architectural support by building Proof of Concepts & Prototypes Self-Starter to deliver data engineering solutions to optimize both the cost and existing solution Requirements for the Senior Data Engineer (Cloud, Azure, SQL, Python, Java) Extensive Software Industry experience Development experience with Azure services (such as Data Factories, Data Lake Storage, SQL Elastic Pools, Data Pipeline, DataBricks, Kubernetes Services, Apache Nifi, jFrog etc) Development experience with GCP services (Cloud Storage, Cloud Spanner, BigQuery) Experience with Apache Spark and NoSQL Implementation Extensive working knowledge in different programming or scripting languages like Scala, Java, Linux, Shell, SQL, Python Proficiency working with structured, semi-structured and unstructured data sets including social, web logs and real time streaming data feeds Able to tune Big Data solutions to improve performance and end-user experience Knowledge on Visualization and Data Science Tools Expert level usage with Jenkins, GitHub is preferred If you would like to apply for the Senior Data Engineer (Cloud, Azure, SQL, Python, Java) role then please click the apply button