Sr. LLM Engineer

Enter the world of artificial intelligence by working with global AI giants and disruptors.

We’re on the lookout for a Sr. LLM Engineer - with strong backend expertise.

As an LLM leader at Acme AI, you'll play a critical role in architecting and implementing backend systems and LLM integrations for powering cloud-based API-driven Retrieval Augmented Generation (RAG) and NL2SQL applications for our international clients.

KEY RESPONSIBILITIES

General responsibilities include:

  • Design and develop Python-based backend services (Fast API / Flask) for LLM based RAG systems and data engines
  • Build and optimise document ingestion, embedding, retrieval and generation pipelines using vector databases (PG Vector, FAISS, Pinecone)
  • Develop and optimise RAG workflows, including document chunking, context retrieval, and response generation
  • Integrate and orchestrate LLM providers(OpenAI, Anthropic, Azure OpenAI) through LangChain or custom orchestration layers
  • Build and optimise APIs for documentation, retrieval and generation
  • Implement NL2SQL pipelines (baseline and agentic) for natural-language-to-query conversion and visualisation suggestion
  • Ensure scalability, reliability, and security of Python services deployed on AWS (ECS, Lambda, S3, RDS) and MS Azure
  • Ensure data privacy, security and compliance across backend systems
  • Monitor performance, latency, and cost of LLM calls, embeddings, and retrieval operations
  • Collaborate closely with the Controller (Node.js) team to align orchestration and response pipelines
  • Lead code reviews, unit testing, and CI/CD automation for Python-based backend components
  • Document APIs and backend workflows, ensuring clarity for frontend and data engineering teams

EXPERIENCE + QUALIFICATION

  • 5–7 years of backend engineering experience with Python (FastAPI, Flask, or Django)
  • Proven experience building LLM-integrated RAG systems or AI-driven backend architectures
  • Strong understanding of vector databases(PGVector, FAISS, Pinecone, Weaviate) and embedding models
  • Hands-on experience with LangChain, LlamaIndex, or similar LLM orchestration frameworks
  • Strong knowledge of AWS services (ECS, Lambda, S3, RDS, CloudWatch) and CI/CD pipelines
  • Familiarity with prompt engineering, token optimisation and multi-model LLM orchestration
  • Knowledge of data processing workflows (CSV/XLS ingestion, schema parsing, SQL query generation)
  • Commitment to best practices in code quality, security, and cloud-native design

NICE TO HAVE

  • Experience with agentic LLM pipelines (multi-step reasoning and tool use)
  • Familiarity with multi-tenant architectures and event-driven systems (SQS, SNS, Kafka)
  • Knowledge of MLOps / LLMOps practices for monitoring and fine-tuning model usage
  • Background in data visualisation APIs or analytics-driven query generation
  • Experience in Microsoft Azure services and technologies

WORK DETAILS

This is a full-time position, where you’ll be working in a supportive, fast-paced environment with access to learning and development opportunities.

  • Type: Full-time, onsite with some remote flexibility
  • Schedule: Standard 8 hours, five-days engagement under contract
  • Location: Acme AI Ltd, Level 5, House 385, Road 6, Mirpur DOHS, Dhaka 1216
  • Salary: BDT 90,000-130,000 (offered based on experience and test results)

BENEFITS

  • Competitive salary based on experience
  • Festival bonuses
  • Access to company health fund (upon eligibility; since health and hospitalisation insurance sucks!)
  • Quarterly profit-sharing (upon eligibility)
  • Acme AI is the largest AI enterprise in Bangladesh - having the largest AI data curation team in the country. Therefore, your developed LLM solutions can benefit from on-demand SFT/RLHF support subject to client approval.

How to apply?

Please lodge your applications by filling out this Google form.

Application Deadline: Considered on a rolling basis until 2 November 2025 at 11.59 PM