City of London, London, United Kingdom Hybrid / WFH Options
Anson McCade
varied use cases. Build agentic workflows and reasoning pipelines using frameworks such as LangChain, LangGraph, CrewAI, Autogen, and LangFlow. Implement retrieval-augmented generation (RAG) pipelines using vector databases like Pinecone, FAISS, Chroma, or PostgreSQL. Fine-tune prompts to optimise performance, reliability, and alignment. Design and implement memory modules for short-term and long-term agent behaviours. Deploy models and orchestrate More ❯
varied use cases. Build agentic workflows and reasoning pipelines using frameworks such as LangChain, LangGraph, CrewAI, Autogen, and LangFlow. Implement retrieval-augmented generation (RAG) pipelines using vector databases like Pinecone, FAISS, Chroma, or PostgreSQL. Fine-tune prompts to optimise performance, reliability, and alignment. Design and implement memory modules for short-term and long-term agent behaviours. Deploy models and orchestrate More ❯
patterns specific to RAG (Retrieval-Augmented Generation), Graph RAG, Agentic RAG, and multi-agent systems. Vector Databases & Embeddings: Expertise in working with various embedding models and vector databases (e.g., Pinecone, Weaviate, Chroma, FAISS). Advanced AI Concepts: Strong grasp of advanced techniques such as complex task decomposition for agents, reasoning engines, knowledge graphs, autonomous agent design, and evaluation methodologies for More ❯
and modern web frameworks Deep experience with AI/ML frameworks (PyTorch, TensorFlow, Transformers, LangChain) Mastery of prompt engineering and fine-tuning Large Language Models Proficient in vector databases (Pinecone, Weaviate, Milvus) and embedding technologies Expert in building RAG (Retrieval-Augmented Generation) systems at scale Strong experience with MLOps practices and model deployment pipelines Proficient in cloud AI services (AWS More ❯
and modern web frameworks Deep experience with AI/ML frameworks (PyTorch, TensorFlow, Transformers, LangChain) Mastery of prompt engineering and fine-tuning Large Language Models Proficient in vector databases (Pinecone, Weaviate, Milvus) and embedding technologies Expert in building RAG (Retrieval-Augmented Generation) systems at scale Strong experience with MLOps practices and model deployment pipelines Proficient in cloud AI services (AWS More ❯
Data Acumen: Solid understanding of data requirements for machine learning models, including feature engineering, data validation, and dataset versioning. Vector Database Experience: Practical experience working with vector databases (e.g., Pinecone, Milvus, Chroma) for embedding storage and retrieval. Generative AI Familiarity: Understanding of data paradigms for LLMs, RAG architectures, and how data pipelines support fine-tuning or pre-training. MLOps Principles More ❯
largescale transformer models (BERT, GPT) and promptengineering for sentiment tasks Background building activelearning and annotation pipelines to bootstrap training data Familiarity with semantic search or vector databases (Elasticsearch, FAISS, Pinecone) for topic modeling and similarity queries Familiarity with crypto markets, order books, and risk-management frameworks Familiarity with anomalydetection methods for streaming text and timeseries data Experience developing EVM smart More ❯
DevOps tools such as GitHub Actions, Azure DevOps, or Terraform. Understanding of search technologies like Elasticsearch, Meilisearch, or Typesense. Experience working with vector search or hybrid search (e.g. pgvector, Pinecone). Knowledge of Microsoft Entra ID/Azure AD and web authentication protocols (OAuth, OpenID Connect). Experience with serverless functions, real-time systems, or edge computing. We are united More ❯
controllers. Develop and maintain AI microservices using Docker, Kubernetes, and FastAPI, ensuring smooth model serving and error handling; Vector Search & Retrieval: Implement retrieval-augmented workflows: ingest documents, index embeddings (Pinecone, FAISS, Weaviate), and build similarity search features. Rapid Prototyping: Create interactive AI demos and proofs-of-concept with Streamlit, Gradio, or Next.js for stakeholder feedback; MLOps & Deployment: Implement CI/… experience fine-tuning LLMs via OpenAI, HuggingFace or similar APIs; Strong proficiency in Python; Deep expertise in prompt engineering and tooling like LangChain or LlamaIndex; Proficiency with vector databases (Pinecone, FAISS, Weaviate) and document embedding pipelines; Proven rapid-prototyping skills using Streamlit or equivalent frameworks for UI demos. Familiarity with containerization (Docker) and at least one orchestration/deployment platform More ❯