Loading...

GOQii

Next Gen AI Engineers

GOQii
Mumbai, Maharashtra On-site
On-Site Full-Time Mumbai, Maharashtra India

About the Role

GOQii is a global health-tech company on a mission to improve health, well-being, and longevity. We combine the power of human coaches, real-time data, and AI to drive lasting behavior change. Our ecosystem spans across preventive health, diagnostics, insurance, and wellness—impacting lives at scale.
Backed by top investors and partners, we're now building the next wave of AI-first products that shape the future of digital health and human performance

About the Role

We are on the hunt for highly skilled and curious AI Engineers to join our next-gen AI product team. You will be working at the bleeding edge of applied AI—developing LLM-powered systems, agentic architectures, and AI-first applications using the best from OpenAI, Google (Gemini), Meta (LLaMA 3), Anthropic (Claude), Mistral, and more.

If you’re excited by RAG, prompt engineering, multi-agent orchestration, fine-tuning, and everything in between — we want to talk to you.

What You’ll Do
Build and fine-tune LLMs (OpenAI GPT-4o, Claude 3, Gemini 1.5, LLaMA 3, Mistral, etc.) for specialized use cases

Architect RAG (Retrieval-Augmented Generation) pipelines with custom vector databases
Design and deploy Agentic Systems using LangChain, CrewAI, AutoGen, and more
Implement multi-agent orchestration with planning, memory, and tool use
Integrate AI with structured systems, APIs (e.g., OpenAI functions, tool calling, ReAct pattern)
Build full-stack AI flows including backend (FastAPI / Node.js), vector DBs, and client-side agents
Deploy models using GCP Vertex AI, AWS Bedrock/Sagemaker, Azure OpenAI, etc.
Collaborate with product and design teams to turn vision into deployable AI-first products

Skills We’re Looking For
Core AI/ML Skills:
Large Language Models: GPT-4o, Claude 3, Gemini, LLaMA 3, Mistral
RAG Architectures: Embeddings, Vector Stores, Hybrid Search
Prompt Engineering: Few-shot, Chain-of-Thought, Self-Refine, Function Calling
Fine-Tuning & LoRA: PEFT, QLoRA, model adapters
Open Source Models: Hugging Face, GGUF, Ollama, vLLM

Agentic Frameworks:
LangChain, CrewAI, AutoGen, Autogen Studio
Semantic Kernel, DSPy
Toolformer-style tool integration (agents that call tools/APIs)
Knowledge Graphs & Reasoning:
LLM + KG integration
Contextual memory + multi-hop reasoning
Knowledge-based planning (Plan & Solve / ReAct)

Infrastructure & Deployment:
Google Vertex AI, AWS Bedrock, Azure AI, RunPod, Modal
FastAPI, Docker, Python AsyncIO
Vector DBs: FAISS, Pinecone, Weaviate, Qdrant

Nice to Have:
Multi-modal experience (e.g., Vision + LLM, Speech-to-Text, Whisper, Gemini multimodal)
Speech AI (TTS, STT, Whisper, OpenAI Voice Engine)
WebRTC or voice agent integration
Real-time or streaming data use cases
AI agents in frontends (Next.js + WebLLM, AgentJS, LangGraph)

You Should Have
2–5 years of experience building real-world AI applications (not just research)
Hands-on experience with LLMs, RAG, fine-tuning, and/or agentic workflows
Strong Python and API integration skills
Excellent understanding of AI model limitations and strengths
Fast learner who can work with minimal supervision in a fast-moving environment

Bonus Points
GitHub/open-source contributions in AI/LLMs
Built your own AI agent or app
Contributions to LangChain, Hugging Face, CrewAI, etc.
Startup experience or founder mindset

Why Join Us?
Work on AI-first products that matter — in health, productivity, and human longevity
Collaborate with an elite product + AI + engineering team
Fast-track your growth with real autonomy and full-stack impact
Work with the latest models and tools as part of your job

How to Apply
Send your profile (CV + GitHub or portfolio) to [email protected] or apply via LinkedIn.

Apply for this position

Log in or Sign up to Apply

Access the application form by logging in or creating an account.

Application Status

Application Draft

In Progress

Submit Application

Pending

Review Process

Expected within 5-7 days

Similar Jobs