In this video, we delve into the challenge of evaluating the quality of AI-generated text, focusing on how to ensure it is accurate and avoids…
Continue reading → Streamlit & Ollama: Querying Mistral LLM Locally and Generating Titles & Keywords
Bruno Flaven's website
In this video, we delve into the challenge of evaluating the quality of AI-generated text, focusing on how to ensure it is accurate and avoids…
Continue reading → Streamlit & Ollama: Querying Mistral LLM Locally and Generating Titles & Keywords
An article exploring the process for testing the output of Large Language Models (LLMs) using a tool called “promptfoo.” This tool allows developers to evaluate…
Continue reading → Promptfoo: The Ultimate Tool for Ensuring LLM Quality and Reliability (Part 2)
An article exploring the process for testing the output of Large Language Models (LLMs) using a tool called “promptfoo.” This tool allows developers to evaluate…
Continue reading → Promptfoo: The Ultimate Tool for Ensuring LLM Quality and Reliability (Part 1)
This video is part of a post “Building a Vue.js SPA (Single Page Application) Frontend with FastAPI Backend for AI Integration” QUICK DESCRIPTION Transitioning from…
Continue reading → Video #6 Streamlit Mastery: View Whisper Transcripts & Prevent Page Reloads
This video is part of a post “Building a Vue.js SPA (Single Page Application) Frontend with FastAPI Backend for AI Integration” QUICK DESCRIPTION Transitioning from…
Continue reading → Video #4 Connect FastAPI Backend to Vue.js Frontend: Full Guide
This video is part of a post “Building a Vue.js SPA (Single Page Application) Frontend with FastAPI Backend for AI Integration” QUICK DESCRIPTION Transitioning from…
Continue reading → Video #1 Scaffold Your First Vitejs App: Comprehensive Beginner’s Guide
This video is part of a post “Empower Your Workflow: Harnessing the Power of LM Studio and Ollama for Seamless Local LLM Execution” POST: https://wp.me/p3Vuhl-3iX…
This video is part of a post “Empower Your Workflow: Harnessing the Power of LM Studio and Ollama for Seamless Local LLM Execution” POST: https://wp.me/p3Vuhl-3iX…