AI Engineer
Job Description
We’re looking for a talented AI Engineer to join our team focused on implementing and scaling large language models (LLMs) and generative AI systems. In this role, you will bridge the gap between cutting-edge research and practical applications, turning innovative AI concepts into robust, efficient, and production-ready systems. You will work closely with our research team and data engineers to build and optimize AI solutions that drive our company's products and services.
Key Responsibilities
- Implement and optimize large language models and generative AI systems for production environments
- Collaborate with researchers to translate research prototypes into scalable, efficient implementations
- Design and develop AI infrastructure components for model training, fine-tuning, and inference
- Optimize AI models for performance, latency, and resource utilization
- Implement systems for model evaluation, monitoring, and continuous improvement
- Develop APIs and integration points for AI services within our product ecosystem
- Troubleshoot complex issues in AI systems and implement solutions
- Contribute to the development of internal tools and frameworks for AI development
- Stay current with emerging techniques in AI engineering and LLM deployment
- Collaborate with data engineers to ensure proper data flow for AI systems
- Implement safety measures, content filtering, and responsible AI practices
Required Skills & Qualifications
- Bachelor's or Master's degree in Computer Science, Engineering, or related technical field
- 3+ years of hands-on experience implementing and optimizing machine learning models
- Strong programming skills in Python and related ML frameworks (PyTorch, TensorFlow)
- Experience with deploying and scaling AI models in production environments
- Familiarity with large language models, transformer architectures, and generative AI
- Knowledge of cloud platforms (AWS, GCP, Azure) and containerization technologies
- Understanding of software engineering best practices (version control, CI/CD, testing)
- Experience with ML engineering tools and platforms (MLflow, Kubeflow, etc.)
- Strong problem-solving skills and attention to detail
- Ability to collaborate effectively in cross-functional teams
Preferred Qualifications
- Experience with fine-tuning and prompt engineering for large language models
- Knowledge of distributed computing and large-scale model training
- Familiarity with model optimization techniques (quantization, pruning, distillation)
- Experience with real-time inference systems and low-latency AI services
- Understanding of AI ethics, bias mitigation, and responsible AI development
- Experience with model serving platforms (TorchServe, TensorFlow Serving, Triton)
- Knowledge of vector databases and similarity search for LLM applications
- Experience with reinforcement learning and RLHF techniques
- Familiarity with front-end technologies for AI application interfaces
Compensation
iGenius offers a competitive compensation structure, including salary, performance-based bonuses, and additional components based on experience. All roles include comprehensive benefits as part of the total compensation package.
About Domyn
Domyn is a company specializing in researching and developing Responsible AI for regulated industries, including financial services, government, and heavy industry. It supports enterprises with proprietary, fully governable solutions, based on a composable AI architecture – including LLMs, AI agents and one of the world’s largest supercomputers.
At the core of Domyn’s product offer is a chip-to-frontend architecture that lets organizations control the entire technological AI stack — from hardware to final application — engineered to guarantee full control, isolation, and security across the entire AI lifecycle.
The backbone of this architecture is a state-of-the-art foundational LLM series designed for advanced reasoning, and optimized to understand each business’s specific language, logic, and context. Available under an open-enterprise license, the full ownership of these models – Domyn Large and Domyn Small – can be permanently transferred and assigned to individual organizations.
Once deployed, they enable customizable agents that act on proprietary data to solve problems and perform specialized tasks. All agents and solutions are managed through a unified platform, with native tools for access management, traceability, and security.
Powering it all, the Colosseum supercomputer (currently in development) – one of the world’s largest – will train and deploy next-generation models (+1T parameters) with unprecedented power thanks to NVIDIA Grace Blackwell Superchips, completing Domyn’s chip-to-frontend framework.
Today, Domyn is one of the leaders in the AI landscape, with a client portfolio that includes Allianz, Intesa Sanpaolo and Fincantieri. It also has forged strategic partnerships with global leaders such as Microsoft, as well as NVIDIA and G42 for the development of the Colosseum supercomputer.
Please review our Privacy Policy here https://bit.ly/2XAy1gj .
Company Information
Location: Not specified
Type: Not specified