Senior Research Engineer - Machine Learning Systems
State University of New York at Buffalo - Georgia Center, Vermont, United States
Work at State University of New York at Buffalo
Overview
- View job
Overview
Position Title:
Senior Research Engineer - Machine Learning Systems Department:
Institute for Artificial Intelligence & Data Sciences UB is seeking a
Senior Research Engineer
with deep expertise in areas such as Artificial Intelligence (AI)/machine learning, AI engineering, AI infrastructure, hybrid cloud computing, and parallel programming with GPUs, to work at the Institute for Artificial Intelligence and Data Science (IAD). As a Senior Research Engineer, you should be experienced in the development of scalable AI systems and deployment of distributed ML models, and will be expected to collaborate with researchers across domains. This role includes opportunities to work with state-of-the-art natural language processing, large language models (LLMs), computer vision models, speech models, time series models, and many other related machine learning models, managing AI infrastructure, and optimizing distributed model training and inference. The IAD at the University at Buffalo is a leader in advancing AI, data science, and computational research. IAD fosters innovation by addressing pressing challenges in areas such as education, health care, robotics, and autonomous vehicles, contributing to the future of the U.S. economy and security. This role offers the opportunity to contribute to UB’s pioneering work at the intersection of AI and data science, collaborating with experts to drive advancements that make a lasting societal impact. Key Responsibilities: Design and develop full-stack end-to-end AI solutions to support IAD’s missions. Collaborate with UB’s various decanal units to understand the requirements of AI solutions and build the solutions accordingly. Lead system architecture discussions and the design of the full-stack solution architecture, including front-end, back-end, and cloud deployment (such as AWS, GCP). Design, maintain, and optimize AI infrastructure to support the test and deployment of AI solutions, including distributed ML training and inference. Under the guidance of faculty members, work with and deploy machine learning models (including LLM models) into both on-prem systems and hybrid clouds, ensuring efficient integration into production environments. Build and maintain scalable, real-time data pipelines and MLOps workflows. Mentor junior engineers to build the AI solutions with robust system performance. Minimum Qualifications: Bachelor’s degree in Computer Science, Engineering, or a related field. 3 years of software engineering (end-to-end full-stack development) and machine learning experience (industry or academia), or an equivalent combination of education and experience. Expertise in PyTorch, TensorFlow, distributed ML training, inference, and MLOps tools. Proficient in AI infrastructure, high-performance computing, cloud platforms (AWS, GCP), and parallel programming with GPUs. Knowledge of cloud technologies such as containerization (Docker, Kubernetes) and big data technologies (Hadoop, Spark). Preferred Qualifications: Master’s degree in Computer Science, Engineering, or a related field. Research project experience and academic publications. Experience with multimodal data and data pipeline creation. Experience in working with and deploying LLM and LVLM models. Strong problem-solving, communication, and collaboration skills. Salary Range: $100,000 - $125,000 Work Hours: 37.5 hours per week As an Equal Opportunity / Affirmative Action employer, the Research Foundation will not discriminate in its employment practices due to an applicant’s race, color, religion, sex, sexual orientation, gender identity, national origin and veteran or disability status.
#J-18808-Ljbffr