Aurelio logo

learning center

At Aurelio AI we believe in the power of online learning. That's why we dedicate a lot of our efforts to developing educational material. Here is some of the latest content and talks that members of the Aurelio team have published.

Fine-Tuning in Sentence Transformers 3

Information Retrieval

Fine-Tuning in Sentence Transformers 3

Fine-Tune Embedding models using Sentence Transformers 3 for better RAG performance.

Getting Started with NVIDIA's AI Workbench

Tooling

Getting Started with NVIDIA's AI Workbench

Everything you need to know about the foundation of NVIDIA's AI offerings, the AI Workbench.

Getting Started with Semantic Chunking

AI Engineering

Getting Started with Semantic Chunking

Everything you need to know about different types of semantic chunking for RAG.

Setup a Remote NVIDIA AI Workbench Node using EC2

Tooling

Setup a Remote NVIDIA AI Workbench Node using EC2

How to run GPU-accelerated ML workloads with ease using NVIDIA's AI Workbench and a CUDA-enabled EC2 instance.

Semantic Chunking of Videos for AI

AI Engineering

Semantic Chunking of Videos for AI

How to intelligent process video for downstream AI applications using semantic chunking.

Mixtral 8X7B Intro

AI Engineering

Mixtral 8X7B Intro

An introduction to the first open-weights model to outperform GPT-3.5 and how to use it as an adbanced AI agent.

Pinecone x LangChain SF Hacking Hours

Events

Pinecone x LangChain SF Hacking Hours

Join James and Harrison at LangChain HQ in San Francisco for a night of hacking with Pinecone and LangChain.

Mitigate Hallucinations Through RAG (AWS Blog)

Information Retrieval

Mitigate Hallucinations Through RAG (AWS Blog)

We explore how to mitigate hallucinations in Large Language Models (LLMs) using Retrieval Augmented Generation (RAG) and the Pinecone vector database in Amazon SageMaker JumpStart.

AI Day Tokyo 2023

Events

AI Day Tokyo 2023

Join James and a packed lineup of experts in AI at AI Day Tokyo in Shibuya on December 2023.

Alternative Embedding Models to OpenAI

Information Retrieval

Alternative Embedding Models to OpenAI

A look at a few of the best retrieval models for Retrieval Augmented Generation (RAG) and how we use them.

RAG and Rerankers with Cohere AI

Information Retrieval

RAG and Rerankers with Cohere AI

James explains how we can use Cohere AI's reranking models to massively enhance the performance of Retrieval Augmented Generation (RAG) pipelines.

Oslo: Large Language Models and Search (Tekna)

Events

Oslo: Large Language Models and Search (Tekna)

Large Language Models (LLMs) are revolutionalizing how machines understand and generate human-like text. In this Tekna meetup, James (Aurelio AI) and Jo Kristian Bergum (Vespa) explain how to get them working with retrieval.

Deploying HuggingFace Models on NVIDIA-Enabled EKS Nodes

AI Engineering

Deploying HuggingFace Models on NVIDIA-Enabled EKS Nodes

Bogdan shares how he used battle-tested container orchestration for our GPU-enabled inference workloads, with baked-in telemetry from Prometheus.

Fine-tuning OpenAI's GPT 3.5

AI Engineering

Fine-tuning OpenAI's GPT 3.5

Fine-tuning for GPT 3.5 turbo is finally here! The latest update gives OpenAI users the ability to create their own custom GPT-3.5 model that has been tuned towards a particular dataset. James explains how it works here.

Nvidia's NeMo Guardrails: Full Walkthrough

AI Engineering

Nvidia's NeMo Guardrails: Full Walkthrough

James explores Nvidia's NeMo Guardrails library for building chatbots. Allowing us to create semi or fully deterministic shields that use against specific behaviors, conversation topics, or even to trigger particular actions.

Superagent.sh August Meetup

Events

Superagent.sh August Meetup

Simonas shares his thoughts on the Superagent.sh library and how it helps us develop agent frameworks for our clients.

Supercharging Llama-2 with RAG

Events

Supercharging Llama-2 with RAG

James and the LangChain team explore how we add external knowledge to Large Language Models using Retrieval Augmented Generation (RAG) using the Pinecone vector database.

Podcast Chatbot using LangChain Agents and GPT-3.5

AI Engineering

Podcast Chatbot using LangChain Agents and GPT-3.5

Conversational agents are set to become the future of natural language interfaces. Learn how they work in this LangChain walkthrough with James.

Retreival Augmented Chat using GPT-4 and Pinecone

AI Engineering

Retreival Augmented Chat using GPT-4 and Pinecone

GPT-4 is the completion model from OpenAI. This walkthrough from James demonstrates how to ground it with real world information using a vector database.

Building the Future with LLMs, LangChain, & Pinecone

AI Engineering

Building the Future with LLMs, LangChain, & Pinecone

Join Harrison Chase (founder of LangChain), and James Briggs, (Aurelio and Pinecone), as we explore the new age of AI and it's biggest enabler — LangChain.

Medical Search Engine with SPLADE + Sentence Transformers

Information Retrieval

Medical Search Engine with SPLADE + Sentence Transformers

A walkthrough from James on hybrid search applied to the Medical domain. Hybrid search is a powerful solution in language-only and multi-modal retrieval systems.