Want to hear how I work? Hit play.Find roles with Kablio AI to help build and power the world.Kablio AI helps you secure roles in construction, clean energy, facilities management, engineering, architecture, sustainability, environment and other physical world sectors.
Get hired, get rewarded!
Land a job through Kablio and earn a 5% salary bonus.
Exclusive benefits
5%Bonus
AI QA Engineer - GenAI
Ge Vernova
Energy tech co. that designs, builds, services gas, nuclear, hydro, steam & wind power systems globally.
Design and lead AI benchmark testing frameworks for GE Vernova.
Design, develop, and execute comprehensive benchmark frameworks to evaluate AI use cases, prompt performance, robustness, across diverse array of use cases.
Monitor AI use cases performance in production, systematically identify issues, and recommend improvements to maintain high standards of AI quality and safety.
Stay current with emerging AI technologies, contribute to platform and tooling improvements, and share knowledge within the team.
Conduct quantitative and qualitative evaluations of AI outputs ensuring alignment with business objectives, regulatory standards, and ethical AI principles.
Document testing methodologies, framework designs, quality metrics, and ensure thorough reporting to stakeholders.
Support continuous improvement initiatives by incorporating feedback loops and new benchmarking techniques as AI technologies evolve.
Develop automated testing suites to validate AI functionalities such as prediction accuracy, response consistency, case handling, bias detection, and model degradation over time.
Collaborate closely with Prompt Engineers, AI Agent Engineers and product owners to define quality standards, acceptance criteria, and key performance indicators (KPIs) for AI deployments.
Provide training and mentorship on AI testing best practices within GE Vernova’s Prompt engineering and AI Agent Engineering team.
What you bring
llms
python
aws
prompt engineering
master’s
qa
Deep understanding of machine learning models, LLMs, and AI validation methodologies.
Familiarity with RAG systems, vector databases, and GenAI architectures.
Strong communication skills with ability to document and present technical information clearly.
Ability to analyze complex datasets and performance metrics quantitatively and qualitatively.
Strong programming skills in Python and experience with automated testing frameworks (e.g., pytest, Selenium).
Knowledge of cloud environments (AWS, Azure, or GCP), containerization, and deployment pipelines.
Hands-on experience with LLMs, prompt engineering, and natural language processing (NLP).
Knowledge of agent orchestration platforms and multi-agent systems (e.g., AutogenAI, LangGraph, MCP protocol).
Experience with AI fairness, bias detection, and responsible AI practices.
3+ years of experience in software quality assurance with at least 2 years focused on AI/ML systems.
Master’s degree in Computer Science, AI Engineering, Data Science, or related field.
Proven track record designing and implementing AI benchmark frameworks or quality assurance strategies.
Benefits
Information not given or found
Training + Development
Information not given or found
Interview process
Information not given or found
Visa Sponsorship
Information not given or found
Security clearance
offer conditioned upon successful completion of a drug screen.
Hey there! Before you dive into all the good stuff on our site, let’s talk cookies—the digital kind. We use these little helpers to give you the best experience we can, remember your preferences, and even suggest things you might love. But don’t worry, we only use them with your permission and handle them with care.