Back to Jobs

Senior AI Engineer

Velsera No location specified Full-time
$120,000
per year

Job Description

Medicine moves too slow. At Velsera, we are changing that.

Velsera was formed in 2023 through the shared vision of Seven Bridges and Pierian, with a mission to accelerate the discovery, development, and delivery of life-changing insights.

Velsera provides software and professional services for:

  • AI-powered multimodal data harmonization and analytics for drug discovery and development
  • IVD development, validation, and regulatory approval
  • Clinical NGS interpretation, reporting, and adoption

With our headquarters in Boston, MA, we are growing and expanding our teams located in different countries!

What will you do?

  • Train, fine-tune, and deploy Large Language Models (LLMs) to solve real-world problems effectively.
  • Design, implement, and optimize AI/ML pipelines to support model development, evaluation, and deployment.
  • Collaborate with Architect, software engineers, and product teams to integrate AI solutions into applications.
  • Ensure model performance, scalability, and efficiency through continuous experimentation and improvements.
  • Work on LLM optimization techniques, including Retrieval-Augmented Generation (RAG), prompt tuning, etc.
  • Manage and automate the infrastructure necessary for AI/ML workloads while keeping the focus on model development.
  • Work with DevOps teams to ensure smooth deployment and monitoring of AI models in production.
  • Stay updated on the latest advancements in AI, LLMs, and deep learning to drive innovation.

What do you bring to the table?

  • Strong experience in training, fine-tuning, and deploying LLMs using frameworks like PyTorch, TensorFlow, or Hugging Face Transformers.
  • Hands-on experience in developing and optimizing AI/ML pipelines, from data preprocessing to model inference.
  • Solid programming skills in Python and familiarity with libraries like NumPy, Pandas, and Scikit-learn.
  • Strong understanding of tokenization, embeddings, and prompt engineering for LLM-based applications.
  • Hands-on experience in building and optimizing RAG pipelines using vector databases (FAISS, Pinecone, Weaviate, or ChromaDB).
  • Experience with cloud-based AI infrastructure (AWS, GCP, or Azure) and containerization technologies (Docker, Kubernetes).
  • Experience in model monitoring, A/B testing, and performance optimization in a production environment.
  • Familiarity with MLOps best practices and tools (Kubeflow, MLflow, or similar).
  • Ability to balance hands-on AI development with necessary infrastructure management.
  • Strong problem-solving skills, teamwork, and a passion for building AI-driven solutions.

Company Information

Location: Charlestown, MA

Type: Hybrid