Role Summary
Stellar Technologies is seeking a Machine Learning Engineer (GenAI) to design, build, and deploy next-generation AI systems combining Large Language Models (LLMs), Retrieval-Augmented Generation (RAG), and agentic AI frameworks.
In this role, you will bridge model development and production engineering — developing scalable AI pipelines, integrating real-time APIs, and ensuring high-performance AI services that power enterprise-grade solutions. You will work at the intersection of machine learning, cloud infrastructure, and applied research, collaborating with top engineers and data scientists to deliver intelligent, production-ready AI capabilities.
Key Responsibilities
-
Develop and optimize AI systems leveraging LLMs, RAG, and agentic AI frameworks (LangChain, LangGraph).
-
Build and deploy production-grade ML pipelines with real-time inference and retrieval components.
-
Design and manage APIs and streaming services to integrate AI models into enterprise platforms.
-
Implement containerized, orchestrated deployments using Docker, Kubernetes, and Azure ML.
-
Automate data preprocessing, model training, evaluation, and versioning pipelines.
-
Collaborate with cross-functional teams to integrate models into front-end, analytics, and automation workflows.
-
Ensure governance, compliance, and security of deployed AI workloads.
-
Conduct performance benchmarking and optimize inference latency and cost.
-
Monitor AI systems in production using observability frameworks (logging, metrics, tracing).
-
Participate in architecture discussions to enhance scalability and reliability of AI services.
Required Skills & Experience
-
Strong hands-on experience with LLMs, RAG, and agentic frameworks (LangChain, LangGraph, Semantic Kernel, etc.).
-
Proficiency in Python, with deep understanding of ML libraries like PyTorch, TensorFlow, scikit-learn, Hugging Face Transformers.
-
Solid experience in API and microservices engineering (FastAPI, Flask).
-
Familiarity with streaming architectures and real-time data handling.
-
Knowledge of cloud platforms (Azure preferred), including Azure AI, Cognitive Services, and ML Ops.
-
Experience with containerization and orchestration (Docker, Kubernetes).
-
Understanding of vector databases (Pinecone, Weaviate, FAISS) and retrieval mechanisms.
-
Experience in CI/CD, model deployment, and production monitoring.
Preferred Skills
-
Exposure to GPU-based inference optimization and serverless deployment.
-
Knowledge of observability and monitoring tools for AI (Prometheus, Grafana, Azure Monitor).
-
Experience in model fine-tuning, prompt engineering, or agentic orchestration.
-
Understanding of AI governance, ethical AI, and data privacy frameworks.
Soft Skills
-
Strong analytical and problem-solving mindset.
-
Excellent collaboration and communication skills.
-
Passion for innovation, experimentation, and applied AI.