Join Red Hat as a Senior Principal Software Engineer (ML/AI Tooling) and help democratize AI with open source. This role is located in the Boston, MA office and is part of the newly formed AI Engineering organization. About the Team Red Hat's AI Innovation team leads frontier AI research and productizes outcomes with open source software. The team bridges cutting edge research and production, translating research papers from conferences such as NeurIPS, ICLR, and ICML into tools that run on the Red Hat AI platform. About the Role You will serve as a strategic technical leader, designing frameworks that enable novel AI algorithms in synthetic data generation, model training, and inference time scaling. In this high impact role you will write code at an expert level, define the technical strategy for the team, and shape Red Hat's open source AI/ML strategy. What You Will Do Strategic Architecture & Leadership Define the technical strategy and architecture for AI/ML tooling and frameworks to work smoothly with platform components. Serve as a Subject Matter Expert, advising leadership on designs that drive community adoption. Mentor and coach Principal and Senior engineers, fostering a culture of technical excellence. Influence the broader open source AI ecosystem by designing frameworks that become community staples. Champion best practices for leveraging AI assisted development tools to accelerate coding, testing, and documentation. Engineering & Execution Lead development of complex projects in data generation, model training, inference time scaling, and new algorithms. Set engineering standards for code reviews, testing frameworks, and system documentation. Evaluate and select appropriate technologies for robust, user friendly developer tools. Drive the design of open source tools to ensure they are accessible, extensible, and community ready. What You Will Bring 10+ years of software development experience, with deep expertise building developer tooling for Machine Learning workflows. Ability to work in the Boston, MA office. Proven track record of maintaining or creating AI/ML projects, preferably popular open source or enterprise scale. Deep knowledge of machine learning frameworks such as PyTorch and TensorFlow. Experience with Large Language Models (LLMs) in tooling, fine tuning, or serving contexts. Ability to integrate the latest AI tooling frameworks and best practices into organizational workflows. Proficiency in a modern backend language (Python, Go, Rust, etc.). Demonstrated leadership of technical strategy for tooling frameworks or platform engineering. Proven success designing software frameworks or APIs with significant adoption. Excellent communication skills to influence Red Hat leadership and shape open source strategy. Track record of mentoring senior engineers and raising the technical bar. Considered a Plus Advanced Python expertise in the ML ecosystem. Experience with state of the art post training techniques. Advanced degree (Master's or PhD) in Machine Learning, NLP, or related field. Experience with Red Hat products. Salary Range: $189,600.00 - $312,730.00 (Actual offer based on qualifications) Pay Transparency Red Hat determines compensation based on location, experience, skills, market value, and internal equity. Annual salary is one component of the package. The position may also be eligible for bonus, commission, or equity. For Remote US locations, the actual salary range may differ. About Red Hat Red Hat is the world's leading provider of enterprise open source software solutions, offering Linux, cloud, container, and Kubernetes technologies. We employ over 40,000 associates worldwide, working flexibly across in office, office flex, and remote environments. Benefits Comprehensive medical, dental, and vision coverage Flexible Spending Account - healthcare and dependent care Health Savings Account - high deductible medical plan Retirement 401(k) with employer match Paid time off and holidays Paid parental leave plans for all new parents Leave benefits including disability, paid family medical leave, and paid military leave Additional benefits: employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more! Note: These benefits apply only to full time, permanent associates in the United States. Inclusion at Red Hat Red Hat's culture embraces transparency, collaboration, and inclusion. We strive to ensure equal opportunities and a welcoming environment for all, valuing diverse ideas and perspectives. Equal Opportunity Policy (EEO) Red Hat is an equal opportunity workplace and affirmative action employer. We review applications without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, disability, marital status, or any other prohibited basis. We do not accept unsolicited resumes or CVs from recruitment agencies. Red Hat is not responsible for any fees related to such submissions. Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance with the online application, email application . General inquiries about application status will not receive a reply.
04/02/2026
Full time
Join Red Hat as a Senior Principal Software Engineer (ML/AI Tooling) and help democratize AI with open source. This role is located in the Boston, MA office and is part of the newly formed AI Engineering organization. About the Team Red Hat's AI Innovation team leads frontier AI research and productizes outcomes with open source software. The team bridges cutting edge research and production, translating research papers from conferences such as NeurIPS, ICLR, and ICML into tools that run on the Red Hat AI platform. About the Role You will serve as a strategic technical leader, designing frameworks that enable novel AI algorithms in synthetic data generation, model training, and inference time scaling. In this high impact role you will write code at an expert level, define the technical strategy for the team, and shape Red Hat's open source AI/ML strategy. What You Will Do Strategic Architecture & Leadership Define the technical strategy and architecture for AI/ML tooling and frameworks to work smoothly with platform components. Serve as a Subject Matter Expert, advising leadership on designs that drive community adoption. Mentor and coach Principal and Senior engineers, fostering a culture of technical excellence. Influence the broader open source AI ecosystem by designing frameworks that become community staples. Champion best practices for leveraging AI assisted development tools to accelerate coding, testing, and documentation. Engineering & Execution Lead development of complex projects in data generation, model training, inference time scaling, and new algorithms. Set engineering standards for code reviews, testing frameworks, and system documentation. Evaluate and select appropriate technologies for robust, user friendly developer tools. Drive the design of open source tools to ensure they are accessible, extensible, and community ready. What You Will Bring 10+ years of software development experience, with deep expertise building developer tooling for Machine Learning workflows. Ability to work in the Boston, MA office. Proven track record of maintaining or creating AI/ML projects, preferably popular open source or enterprise scale. Deep knowledge of machine learning frameworks such as PyTorch and TensorFlow. Experience with Large Language Models (LLMs) in tooling, fine tuning, or serving contexts. Ability to integrate the latest AI tooling frameworks and best practices into organizational workflows. Proficiency in a modern backend language (Python, Go, Rust, etc.). Demonstrated leadership of technical strategy for tooling frameworks or platform engineering. Proven success designing software frameworks or APIs with significant adoption. Excellent communication skills to influence Red Hat leadership and shape open source strategy. Track record of mentoring senior engineers and raising the technical bar. Considered a Plus Advanced Python expertise in the ML ecosystem. Experience with state of the art post training techniques. Advanced degree (Master's or PhD) in Machine Learning, NLP, or related field. Experience with Red Hat products. Salary Range: $189,600.00 - $312,730.00 (Actual offer based on qualifications) Pay Transparency Red Hat determines compensation based on location, experience, skills, market value, and internal equity. Annual salary is one component of the package. The position may also be eligible for bonus, commission, or equity. For Remote US locations, the actual salary range may differ. About Red Hat Red Hat is the world's leading provider of enterprise open source software solutions, offering Linux, cloud, container, and Kubernetes technologies. We employ over 40,000 associates worldwide, working flexibly across in office, office flex, and remote environments. Benefits Comprehensive medical, dental, and vision coverage Flexible Spending Account - healthcare and dependent care Health Savings Account - high deductible medical plan Retirement 401(k) with employer match Paid time off and holidays Paid parental leave plans for all new parents Leave benefits including disability, paid family medical leave, and paid military leave Additional benefits: employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more! Note: These benefits apply only to full time, permanent associates in the United States. Inclusion at Red Hat Red Hat's culture embraces transparency, collaboration, and inclusion. We strive to ensure equal opportunities and a welcoming environment for all, valuing diverse ideas and perspectives. Equal Opportunity Policy (EEO) Red Hat is an equal opportunity workplace and affirmative action employer. We review applications without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, disability, marital status, or any other prohibited basis. We do not accept unsolicited resumes or CVs from recruitment agencies. Red Hat is not responsible for any fees related to such submissions. Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance with the online application, email application . General inquiries about application status will not receive a reply.
Principal Machine Learning Engineer, Distributed vLLM Inference Join to apply for the Principal Machine Learning Engineer, Distributed vLLM Inference role at Red Hat Job Summary At Red Hat we believe the future of AI is open and we are on a mission to bring the power of open source LLMs and vLLM to every enterprise. Red Hat Inference team accelerates AI for the enterprise and brings operational simplicity to GenAI deployments. As leading developers, maintainers of the vLLM and LLM D projects, and inventors of state of the art techniques for model quantization and sparsification, our team provides a stable platform for enterprises to build, optimize, and scale LLM deployments. Join us in shaping the future of AI! What You Will Do Develop and maintain distributed inference infrastructure leveraging Kubernetes APIs, operators, and the Gateway Inference Extension API for scalable LLM deployments. Create system components in Go and/or Rust to integrate with the vLLM project and manage distributed inference workloads. Design and implement KV cache aware routing and scoring algorithms to optimize memory utilization and request distribution in large scale inference deployments. Enhance the resource utilization, fault tolerance, and stability of the inference stack. Contribute to the design, development, and testing of various inference optimization algorithms. Actively participate in technical design discussions and propose innovative solutions to complex challenges. Provide timely and constructive code reviews. Mentor and guide fellow engineers, fostering a culture of continuous learning and innovation. What You Will Bring Strong proficiency in Python, GoLang and at least one of the following: Rust, or C++. Experience with cloud native Kubernetes service mesh technologies/stacks such as Istio, Cilium, Envoy (WASM filters), and CNI. A solid understanding of Layer 7 networking, HTTP/2, gRPC, and the fundamentals of API gateways and reverse proxies. Working knowledge of high performance networking protocols and technologies including UCX, RoCE, InfiniBand, and RDMA is a plus. Excellent communication skills, capable of interacting effectively with both technical and non technical team members. A Bachelor's or Master's degree in computer science, computer engineering, or a related field. Preferred Qualifications Experience with the Kubernetes ecosystem, including core concepts, custom APIs, operators, and the Gateway API inference extension for GenAI workloads. Experience with GPU performance benchmarking and profiling tools like NVIDIA Nsight or distributed tracing libraries/techniques like OpenTelemetry. Ph.D. in an ML related domain is a significant advantage. The salary range for this position is $189,600.00 - $312,730.00. Actual offer will be based on your qualifications. Benefits Comprehensive medical, dental, and vision coverage Flexible Spending Account - healthcare and dependent care Health Savings Account - high deductible medical plan Retirement 401(k) with employer match Paid time off and holidays Paid parental leave plans for all new parents Leave benefits including disability, paid family medical leave, and paid military leave Additional benefits including employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more! Equal Opportunity Policy (EEO) Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, marital status, or any other basis prohibited by law. Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance completing our online job application, email .
04/02/2026
Full time
Principal Machine Learning Engineer, Distributed vLLM Inference Join to apply for the Principal Machine Learning Engineer, Distributed vLLM Inference role at Red Hat Job Summary At Red Hat we believe the future of AI is open and we are on a mission to bring the power of open source LLMs and vLLM to every enterprise. Red Hat Inference team accelerates AI for the enterprise and brings operational simplicity to GenAI deployments. As leading developers, maintainers of the vLLM and LLM D projects, and inventors of state of the art techniques for model quantization and sparsification, our team provides a stable platform for enterprises to build, optimize, and scale LLM deployments. Join us in shaping the future of AI! What You Will Do Develop and maintain distributed inference infrastructure leveraging Kubernetes APIs, operators, and the Gateway Inference Extension API for scalable LLM deployments. Create system components in Go and/or Rust to integrate with the vLLM project and manage distributed inference workloads. Design and implement KV cache aware routing and scoring algorithms to optimize memory utilization and request distribution in large scale inference deployments. Enhance the resource utilization, fault tolerance, and stability of the inference stack. Contribute to the design, development, and testing of various inference optimization algorithms. Actively participate in technical design discussions and propose innovative solutions to complex challenges. Provide timely and constructive code reviews. Mentor and guide fellow engineers, fostering a culture of continuous learning and innovation. What You Will Bring Strong proficiency in Python, GoLang and at least one of the following: Rust, or C++. Experience with cloud native Kubernetes service mesh technologies/stacks such as Istio, Cilium, Envoy (WASM filters), and CNI. A solid understanding of Layer 7 networking, HTTP/2, gRPC, and the fundamentals of API gateways and reverse proxies. Working knowledge of high performance networking protocols and technologies including UCX, RoCE, InfiniBand, and RDMA is a plus. Excellent communication skills, capable of interacting effectively with both technical and non technical team members. A Bachelor's or Master's degree in computer science, computer engineering, or a related field. Preferred Qualifications Experience with the Kubernetes ecosystem, including core concepts, custom APIs, operators, and the Gateway API inference extension for GenAI workloads. Experience with GPU performance benchmarking and profiling tools like NVIDIA Nsight or distributed tracing libraries/techniques like OpenTelemetry. Ph.D. in an ML related domain is a significant advantage. The salary range for this position is $189,600.00 - $312,730.00. Actual offer will be based on your qualifications. Benefits Comprehensive medical, dental, and vision coverage Flexible Spending Account - healthcare and dependent care Health Savings Account - high deductible medical plan Retirement 401(k) with employer match Paid time off and holidays Paid parental leave plans for all new parents Leave benefits including disability, paid family medical leave, and paid military leave Additional benefits including employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more! Equal Opportunity Policy (EEO) Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, marital status, or any other basis prohibited by law. Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance completing our online job application, email .
A leading open-source solution provider in Boston seeks a Machine Learning Engineer focused on distributed vLLM infrastructure. Candidates will contribute to developing innovative solutions, maintaining distributed inference infrastructures, and optimizing performance within cloud-native environments. A strong proficiency in Python and Go, along with experience in Kubernetes, is essential. This role offers a salary range of $133,650 - $220,680 based on qualifications and experience, alongside comprehensive benefits.
04/02/2026
Full time
A leading open-source solution provider in Boston seeks a Machine Learning Engineer focused on distributed vLLM infrastructure. Candidates will contribute to developing innovative solutions, maintaining distributed inference infrastructures, and optimizing performance within cloud-native environments. A strong proficiency in Python and Go, along with experience in Kubernetes, is essential. This role offers a salary range of $133,650 - $220,680 based on qualifications and experience, alongside comprehensive benefits.
Job Summary At Red Hat we believe the future of AI is open and we are on a mission to bring the power of open-source LLMs and vLLM to every enterprise. The Red Hat AI Inference team accelerates AI for the enterprise and brings operational simplicity to GenAI deployments. As leading developers, maintainers of the vLLM project, and inventors of state of the art techniques for model quantization and sparsification, our team provides a stable platform for enterprises to build, optimize, and scale LLM deployments. As a Machine Learning Engineer focused on model optimization algorithms, you will work closely with our product and research teams to develop state of the art deep learning software. You will collaborate with technical and research teams to develop LLM training and deployment pipelines, implement model compression algorithms, and productize deep learning research. This role is for those who enjoy bridging research and production, optimizing large models, and contributing to open source AI tooling. Responsibilities Contribute to the design, development, and testing of various inference optimization algorithms in the LLM compressor, Speculators, and vLLM projects. Design, implement, and optimize model compression pipelines using techniques such as quantization and pruning. Develop and maintain speculative decoding frameworks to improve inference speed while maintaining model accuracy. Collaborate closely with research scientists to translate experimental ideas into robust, production ready systems. Profile and optimize end to end LLM performance, including memory usage, latency, and throughput. Benchmark, evaluate, and implement strategies for optimal performance on target hardware. Build tools to streamline model training, evaluation, and deployment. Participate in technical design discussions and propose innovative solutions to complex problems. Contribute to open source projects, code reviews, and documentation; collaborate with internal and external contributors. Mentor and guide team members, fostering a culture of continuous learning and innovation. Stay current with LLM architectures, inference optimizations, quantization research, and CPU/GPU hardware advancements. Qualifications Strong understanding of machine learning and deep learning fundamentals with experience in one or more of LLM inference optimizations and NLP. Experience with tensor math libraries such as PyTorch and NumPy. Strong programming skills with proven experience implementing Python based machine learning solutions. Ability to develop and implement research ideas and algorithms. Experience with mathematical software, especially linear algebra. Understanding of linear algebra, gradients, probability, and graph theory. Strong communication skills with both technical and non technical team members. BS or MS in computer science, computer engineering, or a related field. A PhD in an ML related domain is considered a strong plus. Benefits Comprehensive medical, dental, and vision coverage; Flexible Spending Account; Health Savings Account; Retirement 401(k) with employer match; Paid time off and holidays; Paid parental leave; Leave benefits including disability, paid family medical leave, and paid military leave; Additional benefits including employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more. (Note: These benefits are only applicable to full time, permanent associates at Red Hat located in the United States.) Inclusion at Red Hat Red Hat's culture is built on the open source principles of transparency, collaboration, and inclusion, where the best ideas can come from anywhere and anyone. When this is realized, it empowers people from different backgrounds, perspectives, and experiences to come together to share ideas, challenge the status quo, and drive innovation. Our aspiration is that everyone experiences this culture with equal opportunity and access, and that all voices are not only heard but also celebrated. We hope you will join our celebration, and we welcome and encourage applicants from all the beautiful dimensions that compose our global village. Equal Opportunity Policy (EEO) Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, medical condition, marital status, or any other basis prohibited by law.
04/02/2026
Full time
Job Summary At Red Hat we believe the future of AI is open and we are on a mission to bring the power of open-source LLMs and vLLM to every enterprise. The Red Hat AI Inference team accelerates AI for the enterprise and brings operational simplicity to GenAI deployments. As leading developers, maintainers of the vLLM project, and inventors of state of the art techniques for model quantization and sparsification, our team provides a stable platform for enterprises to build, optimize, and scale LLM deployments. As a Machine Learning Engineer focused on model optimization algorithms, you will work closely with our product and research teams to develop state of the art deep learning software. You will collaborate with technical and research teams to develop LLM training and deployment pipelines, implement model compression algorithms, and productize deep learning research. This role is for those who enjoy bridging research and production, optimizing large models, and contributing to open source AI tooling. Responsibilities Contribute to the design, development, and testing of various inference optimization algorithms in the LLM compressor, Speculators, and vLLM projects. Design, implement, and optimize model compression pipelines using techniques such as quantization and pruning. Develop and maintain speculative decoding frameworks to improve inference speed while maintaining model accuracy. Collaborate closely with research scientists to translate experimental ideas into robust, production ready systems. Profile and optimize end to end LLM performance, including memory usage, latency, and throughput. Benchmark, evaluate, and implement strategies for optimal performance on target hardware. Build tools to streamline model training, evaluation, and deployment. Participate in technical design discussions and propose innovative solutions to complex problems. Contribute to open source projects, code reviews, and documentation; collaborate with internal and external contributors. Mentor and guide team members, fostering a culture of continuous learning and innovation. Stay current with LLM architectures, inference optimizations, quantization research, and CPU/GPU hardware advancements. Qualifications Strong understanding of machine learning and deep learning fundamentals with experience in one or more of LLM inference optimizations and NLP. Experience with tensor math libraries such as PyTorch and NumPy. Strong programming skills with proven experience implementing Python based machine learning solutions. Ability to develop and implement research ideas and algorithms. Experience with mathematical software, especially linear algebra. Understanding of linear algebra, gradients, probability, and graph theory. Strong communication skills with both technical and non technical team members. BS or MS in computer science, computer engineering, or a related field. A PhD in an ML related domain is considered a strong plus. Benefits Comprehensive medical, dental, and vision coverage; Flexible Spending Account; Health Savings Account; Retirement 401(k) with employer match; Paid time off and holidays; Paid parental leave; Leave benefits including disability, paid family medical leave, and paid military leave; Additional benefits including employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more. (Note: These benefits are only applicable to full time, permanent associates at Red Hat located in the United States.) Inclusion at Red Hat Red Hat's culture is built on the open source principles of transparency, collaboration, and inclusion, where the best ideas can come from anywhere and anyone. When this is realized, it empowers people from different backgrounds, perspectives, and experiences to come together to share ideas, challenge the status quo, and drive innovation. Our aspiration is that everyone experiences this culture with equal opportunity and access, and that all voices are not only heard but also celebrated. We hope you will join our celebration, and we welcome and encourage applicants from all the beautiful dimensions that compose our global village. Equal Opportunity Policy (EEO) Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, medical condition, marital status, or any other basis prohibited by law.
Overview At Red Hat, we connect an innovative community of customers, partners, and contributors to deliver an open source stack of trusted, high-performing solutions. We offer cloud, Linux, middleware, virtualization, and AI technologies, together with award-winning global customer support, consulting, and implementation services. Red Hat is a rapidly-growing company supporting more than 90% of Fortune 500 companies. Red Hat's Global Engineering team is looking for a Senior Software Engineer to join the Agentic and AI Engineering Tools team to develop a Kubernetes/OpenShift Operator and associated software for orchestrating deployment of Llama Stack. In this role, you'll contribute directly to Red Hat's rapidly growing AI/ML family of products and will be responsible for the design, development, and refinement of software adding features that enables Llama Stack to achieve Enterprise readiness. The ideal candidate will have a proven background in developing robust and scalable code. As part of your responsibilities, you will need to adhere to coding best practices and standards, including well-documented, and efficient code; building and implementing upstream unit and E2E automated tests, maintaining updated code documentation and comments, following security best practices, participating in code reviews and other peer review in upstream projects, and staying up-to-date with software engineering technologies, frameworks, and methodologies. What You Will Do Architect and develop a platform for Agentic AI applications. Collaborate with Staff Engineers, Engineering, Product Management, and User Experience to define customer needs and use cases. Collaborate with Quality Engineers to develop and implement comprehensive unit, integration, and end-to-end tests to guarantee the reliability and performance in the upstream project, maintaining CI/CD workflows in GitHub, and ensuring downstream quality. Participate in AI-assisted code reviews, utilizing tools that provide real-time feedback, identify potential bugs, security vulnerabilities, and adherence to coding standards, contributing to a more thorough and efficient review process. Proactively utilize AI-assisted development tools (e.g., GitHub Copilot, Cursor, Claude Code) for code generation, auto-completion, and intelligent suggestions to accelerate development cycles and enhance code quality. Create and maintain clear, concise upstream technical documentation including API references and user guides and collaborating with our internal tech writers to create robust downstream documentation. What You Will Bring 7 to 10 years of advanced Python development experience Advanced knowledge designing robust and scalable software used in highly scaled and performant Distributed Systems Experience with building agents, agentic workflows, or developing with LLMs Knowledge of Kubernetes/OpenShift and operational knowledge building/deploying containers. Experience creating automation for GitHub, using GitHub Actions or related continuous integration tools. Experience developing, deploying or maintaining On-prem or Cloud Infrastructure Advanced knowledge developing unit, functional, and end-to-end (E2E) test cases and automation Ability to quickly learn and use new tools and technologies. The following is considered a plus Experience with open source projects. Experience with Security, Observability, Performance or Scale. Understanding of DevOps methodology, scrum, and/or Jira. Experience with AI and Machine Learning platforms, tools, and frameworks, such as LlamaStack, LangChain, PyTorch, LLaMA.cpp, vLLM, LangGraph, and Kubeflow. Bachelors or Masters degree in computer science or related discipline. Salary and Benefits The salary range for this position is $133,650.00 - $220,680.00. Actual offer will be based on your qualifications. Pay Transparency Red Hat determines compensation based on several factors including but not limited to job location, experience, applicable skills and training, external market value, and internal pay equity. Annual salary is one component of Red Hat's compensation package. This position may also be eligible for bonus, commission, and/or equity. For positions with Remote-US locations, the actual salary range for the position may differ based on location but will be commensurate with job duties and relevant work experience. About Red Hat Red Hat is the world's leading provider of enterprise open source software solutions, using a community-powered approach to deliver high-performing Linux, cloud, container, and Kubernetes technologies. Spread across 40+ countries, our associates work flexibly across work environments, from in-office, to office-flex, to fully remote, depending on the requirements of their role. Red Hatters are encouraged to bring their best ideas, no matter their title or tenure. We're a leader in open source because of our open and inclusive environment. We hire creative, passionate people ready to contribute their ideas, help solve complex problems, and make an impact. Benefits Comprehensive medical, dental, and vision coverage Flexible Spending Account - healthcare and dependent care Health Savings Account - high deductible medical plan Retirement 401(k) with employer match Paid time off and holidays Paid parental leave plans for all new parents Leave benefits including disability, paid family medical leave, and paid military leave Additional benefits including employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more! Note These benefits are only applicable to full time, permanent associates at Red Hat located in the United States. Inclusion and Equal Opportunity Red Hat's culture is built on the open source principles of transparency, collaboration, and inclusion, where the best ideas can come from anywhere and anyone. Our aspiration is that everyone experiences this culture with equal opportunity and access, and that all voices are not only heard but also celebrated. We welcome applicants from all the dimensions that compose our global village. Equal Opportunity Policy (EEO) Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, medical condition, marital status, or any other basis prohibited by law. Red Hat does not seek or accept unsolicited resumes or CVs from recruitment agencies. We are not responsible for, and will not pay, any fees, commissions, or any other payment related to unsolicited resumes or CVs except as required in a written contract between Red Hat and the recruitment agency or party requesting payment of a fee. Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance completing our online job application, email . General inquiries, such as those regarding the status of a job application, will not receive a reply.
04/02/2026
Full time
Overview At Red Hat, we connect an innovative community of customers, partners, and contributors to deliver an open source stack of trusted, high-performing solutions. We offer cloud, Linux, middleware, virtualization, and AI technologies, together with award-winning global customer support, consulting, and implementation services. Red Hat is a rapidly-growing company supporting more than 90% of Fortune 500 companies. Red Hat's Global Engineering team is looking for a Senior Software Engineer to join the Agentic and AI Engineering Tools team to develop a Kubernetes/OpenShift Operator and associated software for orchestrating deployment of Llama Stack. In this role, you'll contribute directly to Red Hat's rapidly growing AI/ML family of products and will be responsible for the design, development, and refinement of software adding features that enables Llama Stack to achieve Enterprise readiness. The ideal candidate will have a proven background in developing robust and scalable code. As part of your responsibilities, you will need to adhere to coding best practices and standards, including well-documented, and efficient code; building and implementing upstream unit and E2E automated tests, maintaining updated code documentation and comments, following security best practices, participating in code reviews and other peer review in upstream projects, and staying up-to-date with software engineering technologies, frameworks, and methodologies. What You Will Do Architect and develop a platform for Agentic AI applications. Collaborate with Staff Engineers, Engineering, Product Management, and User Experience to define customer needs and use cases. Collaborate with Quality Engineers to develop and implement comprehensive unit, integration, and end-to-end tests to guarantee the reliability and performance in the upstream project, maintaining CI/CD workflows in GitHub, and ensuring downstream quality. Participate in AI-assisted code reviews, utilizing tools that provide real-time feedback, identify potential bugs, security vulnerabilities, and adherence to coding standards, contributing to a more thorough and efficient review process. Proactively utilize AI-assisted development tools (e.g., GitHub Copilot, Cursor, Claude Code) for code generation, auto-completion, and intelligent suggestions to accelerate development cycles and enhance code quality. Create and maintain clear, concise upstream technical documentation including API references and user guides and collaborating with our internal tech writers to create robust downstream documentation. What You Will Bring 7 to 10 years of advanced Python development experience Advanced knowledge designing robust and scalable software used in highly scaled and performant Distributed Systems Experience with building agents, agentic workflows, or developing with LLMs Knowledge of Kubernetes/OpenShift and operational knowledge building/deploying containers. Experience creating automation for GitHub, using GitHub Actions or related continuous integration tools. Experience developing, deploying or maintaining On-prem or Cloud Infrastructure Advanced knowledge developing unit, functional, and end-to-end (E2E) test cases and automation Ability to quickly learn and use new tools and technologies. The following is considered a plus Experience with open source projects. Experience with Security, Observability, Performance or Scale. Understanding of DevOps methodology, scrum, and/or Jira. Experience with AI and Machine Learning platforms, tools, and frameworks, such as LlamaStack, LangChain, PyTorch, LLaMA.cpp, vLLM, LangGraph, and Kubeflow. Bachelors or Masters degree in computer science or related discipline. Salary and Benefits The salary range for this position is $133,650.00 - $220,680.00. Actual offer will be based on your qualifications. Pay Transparency Red Hat determines compensation based on several factors including but not limited to job location, experience, applicable skills and training, external market value, and internal pay equity. Annual salary is one component of Red Hat's compensation package. This position may also be eligible for bonus, commission, and/or equity. For positions with Remote-US locations, the actual salary range for the position may differ based on location but will be commensurate with job duties and relevant work experience. About Red Hat Red Hat is the world's leading provider of enterprise open source software solutions, using a community-powered approach to deliver high-performing Linux, cloud, container, and Kubernetes technologies. Spread across 40+ countries, our associates work flexibly across work environments, from in-office, to office-flex, to fully remote, depending on the requirements of their role. Red Hatters are encouraged to bring their best ideas, no matter their title or tenure. We're a leader in open source because of our open and inclusive environment. We hire creative, passionate people ready to contribute their ideas, help solve complex problems, and make an impact. Benefits Comprehensive medical, dental, and vision coverage Flexible Spending Account - healthcare and dependent care Health Savings Account - high deductible medical plan Retirement 401(k) with employer match Paid time off and holidays Paid parental leave plans for all new parents Leave benefits including disability, paid family medical leave, and paid military leave Additional benefits including employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more! Note These benefits are only applicable to full time, permanent associates at Red Hat located in the United States. Inclusion and Equal Opportunity Red Hat's culture is built on the open source principles of transparency, collaboration, and inclusion, where the best ideas can come from anywhere and anyone. Our aspiration is that everyone experiences this culture with equal opportunity and access, and that all voices are not only heard but also celebrated. We welcome applicants from all the dimensions that compose our global village. Equal Opportunity Policy (EEO) Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, medical condition, marital status, or any other basis prohibited by law. Red Hat does not seek or accept unsolicited resumes or CVs from recruitment agencies. We are not responsible for, and will not pay, any fees, commissions, or any other payment related to unsolicited resumes or CVs except as required in a written contract between Red Hat and the recruitment agency or party requesting payment of a fee. Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance completing our online job application, email . General inquiries, such as those regarding the status of a job application, will not receive a reply.
A leading open source software company in Boston seeks a Senior Software Engineer to develop a Kubernetes/OpenShift Operator for the Llama Stack. You will produce robust software while collaborating closely with engineering and product teams. The ideal candidate has at least 7 years of Python development experience, familiarity with Kubernetes/OpenShift, and a strong grasp of test automation. This position offers a competitive salary with significant benefits.
04/02/2026
Full time
A leading open source software company in Boston seeks a Senior Software Engineer to develop a Kubernetes/OpenShift Operator for the Llama Stack. You will produce robust software while collaborating closely with engineering and product teams. The ideal candidate has at least 7 years of Python development experience, familiarity with Kubernetes/OpenShift, and a strong grasp of test automation. This position offers a competitive salary with significant benefits.