AI/ML Developer

0



Codem is a technology services company specializing in eCommerce, SAP, custom applications, cloud infrastructure, DevOps, and systems integration. We work with global enterprises to build and modernize scalable platforms. We are now seeking AI.ML Develper.


ONLY APPLY IF YOU ARE AN IMMEDIATE JOINER.


About the role

We’re looking for an AI Developer to build and ship LLM-powered features (chat/search/agents, RAG pipelines, automations). You’ll work closely with product and data teams to turn messy real-world data into reliable, low-latency experiences.


Responsibilities

Must-have qualifications

  • 3+ years software experience (ideally Python) delivering production code.
  • Hands-on with LLM APIs (OpenAI/Azure OpenAI, Anthropic, or local LLMs like Llama) including prompting, tools/function calling, and streaming.
  • Practical RAG experience using vector databases (e.g., Pinecone, Weaviate, FAISS, pgvector) and embedding models.
  • Experience with LangChain or LlamaIndex (or equivalent in-house orchestration).
  • Strong with web APIs (FastAPI/Flask/Node), Git, testing, and debugging.
  • Solid understanding of security & privacy basics (PII handling, secrets, auth).

Nice to have

  • Reranking (Cohere/TEI), hybrid search (BM25 + embeddings), or Elasticsearch/OpenSearch.
  • Eval frameworks (Ragas, TruLens) and telemetry (Langfuse, OpenTelemetry).
  • Workflow/orchestration (Celery/Temporal/Airflow) and message queues (SQS/Kafka).
  • Cloud: AWS (Bedrock, Lambda), GCP (Vertex AI), Azure (AOAI), Docker; basic Terraform.
  • Frontend collaboration (React) for chat UIs, streaming tokens, and citations.
  • Fine-tuning/LoRA, prompt caching, distillation, or model hosting experience.

Tools you might use here

  • Python (FastAPI), TypeScript/Node (optional), LangChain/LlamaIndex
  • Vector DBs: Pinecone, Weaviate, pgvector/FAISS
  • LLMs/Embeddings: GPT-4/4o/mini, Claude, Llama, instructor/sentence-transformers
  • Infra: AWS/GCP/Azure, Docker, GitHub Actions, Terraform (basic)
  • Obs & Eval: Langfuse, Ragas/TruLens, Prometheus/Grafana

Success in 3–6 months

  • Ship a production RAG feature with measurable uplift in answer quality.
  • Reduce latency/cost via caching/batching and better retrieval configs.
  • Establish evaluation + feedback loop with clear QA dashboards and guardrails.


You have to wait 20 seconds

Generating Apply Link...

Post a Comment

0 Comments
* Please Don't Spam Here. All the Comments are Reviewed by Admin.
Post a Comment (0)
Our website uses cookies to enhance your experience. Learn More
Accept !
X

Join Our WhatsApp Channel to get latest Updates Join Now

Link Copied