learning center
At Aurelio AI we believe in the power of online learning. That's why we dedicate a lot of our efforts to developing educational material. Here is some of the latest content and talks that members of the Aurelio team have published.
![Getting Started with NVIDIA's AI Workbench](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-cloud2.png&w=3840&q=100)
AI
Getting Started with NVIDIA's AI Workbench
Everything you need to know about the foundation of NVIDIA's AI offerings, the AI Workbench.
![Getting Started with Semantic Chunking](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-hand2.png&w=3840&q=100)
Semantic Chunkers
Getting Started with Semantic Chunking
Everything you need to know about different types of semantic chunking for RAG.
![Setup a Remote NVIDIA AI Workbench Node using EC2](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-brain2.png&w=3840&q=100)
Setup a Remote NVIDIA AI Workbench Node using EC2
How to run GPU-accelerated ML workloads with ease using NVIDIA's AI Workbench and a CUDA-enabled EC2 instance.
![Semantic Chunking of Videos for AI](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-brain3.png&w=3840&q=100)
Semantic Chunkers
Semantic Chunking of Videos for AI
How to intelligent process video for downstream AI applications using semantic chunking.
![Mixtral 8X7B Intro](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-hand1.png&w=3840&q=100)
LLMs
Mixtral 8X7B Intro
An introduction to the first open-weights model to outperform GPT-3.5 and how to use it as an adbanced AI agent.
![Pinecone x LangChain SF Hacking Hours](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-code1.png&w=3840&q=100)
Events
Pinecone x LangChain SF Hacking Hours
Join James and Harrison at LangChain HQ in San Francisco for a night of hacking with Pinecone and LangChain.
![Mitigate Hallucinations Through RAG (AWS Blog)](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-landscape3.png&w=3840&q=100)
RAG
Mitigate Hallucinations Through RAG (AWS Blog)
We explore how to mitigate hallucinations in Large Language Models (LLMs) using Retrieval Augmented Generation (RAG) and the Pinecone vector database in Amazon SageMaker JumpStart.
![AI Day Tokyo 2023](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-cloud1.png&w=3840&q=100)
Events
AI Day Tokyo 2023
Join James and a packed lineup of experts in AI at AI Day Tokyo in Shibuya on December 2023.
![Alternative Embedding Models to OpenAI](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-brain1.png&w=3840&q=100)
RAG
Alternative Embedding Models to OpenAI
A look at a few of the best retrieval models for Retrieval Augmented Generation (RAG) and how we use them.
![RAG and Rerankers with Cohere AI](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-information1.png&w=3840&q=100)
RAG
RAG and Rerankers with Cohere AI
James explains how we can use Cohere AI's reranking models to massively enhance the performance of Retrieval Augmented Generation (RAG) pipelines.
![Oslo: Large Language Models and Search (Tekna)](/_next/image?url=%2Fimages%2Fnews%2Fthumbnail-cloud4.png&w=3840&q=100)
Events
Oslo: Large Language Models and Search (Tekna)
Large Language Models (LLMs) are revolutionalizing how machines understand and generate human-like text. In this Tekna meetup, James (Aurelio AI) and Jo Kristian Bergum (Vespa) explain how to get them working with retrieval.
![Deploying HuggingFace Models on NVIDIA-Enabled EKS Nodes](/_next/image?url=%2Fimages%2Fnews%2Farticle-bogdan-hf-eks.jpg&w=3840&q=100)
MLOps
Deploying HuggingFace Models on NVIDIA-Enabled EKS Nodes
Bogdan shares how he used battle-tested container orchestration for our GPU-enabled inference workloads, with baked-in telemetry from Prometheus.
![Fine-tuning OpenAI's GPT 3.5](/_next/image?url=%2Fimages%2Fnews%2Fyoutube-fine-tune-gpt-3.5.jpg&w=3840&q=100)
LLMs
Fine-tuning OpenAI's GPT 3.5
Fine-tuning for GPT 3.5 turbo is finally here! The latest update gives OpenAI users the ability to create their own custom GPT-3.5 model that has been tuned towards a particular dataset. James explains how it works here.
![Nvidia's NeMo Guardrails: Full Walkthrough](/_next/image?url=%2Fimages%2Fnews%2Fyoutube-nemo-guardrails.jpg&w=3840&q=100)
Conversational AI
Nvidia's NeMo Guardrails: Full Walkthrough
James explores Nvidia's NeMo Guardrails library for building chatbots. Allowing us to create semi or fully deterministic shields that use against specific behaviors, conversation topics, or even to trigger particular actions.
![Superagent.sh August Meetup](/_next/image?url=%2Fimages%2Fnews%2Fninja-negan-scofield.jpg&w=3840&q=100)
AI Agents
Superagent.sh August Meetup
Simonas shares his thoughts on the Superagent.sh library and how it helps us develop agent frameworks for our clients.
![Supercharging Llama-2 with RAG](/_next/image?url=%2Fimages%2Fnews%2Flangchain-llama-2.jpg&w=3840&q=100)
LLMs
Supercharging Llama-2 with RAG
James and the LangChain team explore how we add external knowledge to Large Language Models using Retrieval Augmented Generation (RAG) using the Pinecone vector database.
![Podcast Chatbot using LangChain Agents and GPT-3.5](/_next/image?url=%2Fimages%2Fnews%2Fyoutube-langchain-ask-lex.jpg&w=3840&q=100)
Conversational AI
Podcast Chatbot using LangChain Agents and GPT-3.5
Conversational agents are set to become the future of natural language interfaces. Learn how they work in this LangChain walkthrough with James.
![Retreival Augmented Chat using GPT-4 and Pinecone](/_next/image?url=%2Fimages%2Fnews%2Fyoutube-gpt4-chat-and-search.jpg&w=3840&q=100)
LLMs
Retreival Augmented Chat using GPT-4 and Pinecone
GPT-4 is the completion model from OpenAI. This walkthrough from James demonstrates how to ground it with real world information using a vector database.
![Building the Future with LLMs, LangChain, & Pinecone](/_next/image?url=%2Fimages%2Fnews%2Fyoutube-langchain-webinar.jpg&w=3840&q=100)
LLMs
Building the Future with LLMs, LangChain, & Pinecone
Join Harrison Chase (founder of LangChain), and James Briggs, (Aurelio and Pinecone), as we explore the new age of AI and it's biggest enabler — LangChain.
![Medical Search Engine with SPLADE + Sentence Transformers](/_next/image?url=%2Fimages%2Fnews%2Fyoutube-hybrid-search-pubmed.jpg&w=3840&q=100)
Vector Search
Medical Search Engine with SPLADE + Sentence Transformers
A walkthrough from James on hybrid search applied to the Medical domain. Hybrid search is a powerful solution in language-only and multi-modal retrieval systems.