UpTrain (YC W23)

UpTrain (YC W23)

Software Development

San Francisco, CA 1,381 followers

Your open-source LLM Evaluation and Monitoring Toolkit

About us

UpTrain helps solve internal needs (evaluation and prompt experimentation) to external ones and helps instil trust in your users. Some of the critical benefits of UpTrain are: - Diverse evaluations for all your needs - Faster and Systematic Experimentation - Automated Regression Testing - Isolates error cases and finds common patterns among them - Enrich existing datasets by capturing different edge cases encountered in production Check out the repo here: https://github.com/uptrain-ai/uptrain

Website
https://uptrain.ai
Industry
Software Development
Company size
2-10 employees
Headquarters
San Francisco, CA
Type
Privately Held

Locations

Employees at UpTrain (YC W23)

Updates

  • View organization page for UpTrain (YC W23), graphic

    1,381 followers

    🚀 Exciting Update for LLM Developers! 🚀 Delighted to announce a new integration between UpTrain and Promptfoo, aimed at enhancing prompt experimentation for LLM developers. What does this mean for you? 🔍 Compare with Ease: Easily compare outputs from different LLM models and prompt versions. 📊 Analyze Performance: Dive into UpTrain's metrics to evaluate performance across experiments. 📈 Visualize Insights: Utilize Promptfoo's dashboards to visualize experiment results. Whether you're fine-tuning a model or exploring new avenues, this integration equips you with the tools to innovate effectively. Ready to elevate your experimentation? Explore the integration today! #AI #MachineLearning #LanguageModels #UpTrain #Promptfoo #Experimentation

    • No alternative text description for this image
  • View organization page for UpTrain (YC W23), graphic

    1,381 followers

    "What's the right prompt for this application?" "How can I improve this prompt?" Most prompt engineers would be able to relate with these questions.  Experimenting with different versions of prompts is tough for sure, especially when you have to compare them around thousands of data points. UpTrain's newly launched dashboards make prompt experimentation quite easy! 🚀 It lets you compare prompt performance based on metrics like relevance and factual accuracy. The best part is, these dashboards are open-source, you can run them locally on your device. Link in comments #UpTrain #PromptExperimentation #AI

  • View organization page for UpTrain (YC W23), graphic

    1,381 followers

    🚀 Latest update in UpTrain! UpTrain can now simulate and evaluate conversations with AI assistants. Simulate Conversations: Easily simulate conversations with AI assistants for different scenarios. Evaluate Conversations: Evaluate the performance of the assistant based on metrics like user satisfaction, factual accuracy, relevance, and many more. Try it out using: https://lnkd.in/g7UqXKY2

    • No alternative text description for this image
  • UpTrain (YC W23) reposted this

    View organization page for UpTrain (YC W23), graphic

    1,381 followers

    Implementing RAG to an LLM application seems easy, but building a fully functional RAG pipeline is a lot more challenging. A lot of factors can go wrong: - The retrieved context is poor. - The context is not getting utilized effectively. - The LLM is hallucinating, generating incorrect information. and a lot more… These challenges can lead to incomplete or inaccurate responses, undermining the reliability of the LLM system. To understand more about the different problems that can occur in RAG and how to solve them, check out our recent blog: https://lnkd.in/gRCZUMy8

    What's Wrong in my RAG Pipeline? - UpTrain AI

    What's Wrong in my RAG Pipeline? - UpTrain AI

    https://blog.uptrain.ai

  • View organization page for UpTrain (YC W23), graphic

    1,381 followers

    Implementing RAG to an LLM application seems easy, but building a fully functional RAG pipeline is a lot more challenging. A lot of factors can go wrong: - The retrieved context is poor. - The context is not getting utilized effectively. - The LLM is hallucinating, generating incorrect information. and a lot more… These challenges can lead to incomplete or inaccurate responses, undermining the reliability of the LLM system. To understand more about the different problems that can occur in RAG and how to solve them, check out our recent blog: https://lnkd.in/gRCZUMy8

    What's Wrong in my RAG Pipeline? - UpTrain AI

    What's Wrong in my RAG Pipeline? - UpTrain AI

    https://blog.uptrain.ai

  • View organization page for UpTrain (YC W23), graphic

    1,381 followers

    🚀 Introducing our new dashboards, designed to enhance your LLM applications evaluation experience: 1️⃣ Evaluate LLM Applications: Use metrics like relevance, factual accuracy, and more to measure the performance of your LLM applications. 2️⃣ Compare Prompts: Easily compare different versions of prompts to choose the best fit for your use case. 3️⃣ Build Your Own Experiments: Create and manage experiments effortlessly. 4️⃣ Set Up Daily Monitoring: Keep track of your progress with daily monitoring graphs, ensuring your LLM applications are always performing at their best. Check out these dashboards here: https://lnkd.in/gaSYt8Ev #UpTrain #LLM #AI #MachineLearning #Dashboards #Productivity

  • View organization page for UpTrain (YC W23), graphic

    1,381 followers

    🚀 We're excited to introduce the latest enhancements to UpTrain: New Integrations: Ollama: Run evaluations using LLM models hosted locally on your system. Langfuse (YC W23): Easily track your LLM applications for latency, cost, and more. Promptfoo: Conduct experiments to compare prompts and models, visualize results on Promptfoo's dashboards. Zeno: Dive deep into your LLM experiments with interactive dashboards. Helicone: Monitor your LLM applications with detailed dashboards. Automatic Failure Case Identification: UpTrain now automatically identifies failure cases, including issues related to poor quality of retrieved context or inadequate utilization of context, among other challenges. Custom Evaluations: Add Python code and define your own evaluations, such as identifying repetition of words in generated content or analyzing other complex patterns! Upgrade to the latest release of UpTrain (v0.6.10.post1) to check out these updates! 🌟

    • No alternative text description for this image
  • UpTrain (YC W23) reposted this

    View profile for Sourabh Agrawal, graphic

    UpTrain (YC W23) | IIT Bombay

    We are excited to announce the LlamaIndex <> UpTrain (YC W23) integration! It’s been months in making, but we wanted to deliver something of real value to our community. Evaluations are not just about computing a final score for your RAG pipeline but getting 𝐚𝐜𝐭𝐢𝐨𝐧𝐚𝐛𝐥𝐞 𝐢𝐧𝐬𝐢𝐠𝐡𝐭𝐬 𝐨𝐧 𝐰𝐡𝐞𝐫𝐞 𝐭𝐡𝐢𝐧𝐠𝐬 𝐚𝐫𝐞 𝐠𝐨𝐢𝐧𝐠 𝐰𝐫𝐨𝐧𝐠 𝐚𝐧𝐝 𝐡𝐨𝐰 𝐭𝐨 𝐢𝐦𝐩𝐫𝐨𝐯𝐞 𝐭𝐡𝐚𝐭 𝐬𝐜𝐨𝐫𝐞. With this integration, you can evaluate all individual components of your RAG pipeline, such as retrieval, reranking, sub-query, etc. and get deep insights into where your LlamaIndex pipelines need improvements, all with a single line of code. At UpTrain, we are building the gold standard of LLM evaluations with high-quality scores that learn your preferences. • Evaluate different aspects of your application with 20+ preconfigured checks • A high degree of customisation allows you to modify eval prompts, choose evaluator LLM or create your own checks. • Experiment with prompts, LLMs, embedding models, RAG modules, etc. • Do root cause analysis to find failure modes and hidden patterns. and finally, • Interactive dashboards to visualise results and do side-by-side comparisons [More coming soon] Check out the blog: https://lnkd.in/gnwkG2nw Check out UpTrain: https://lnkd.in/gcbM2kY3 It was great fun collaborating with the LlamaIndex team - Ravi Theja Desetty Laurie Voss Jerry Liu! Shikha Mohanty Dhruv Chawla Shreyansh Jain @Ashish Anand Raghav

    Supercharge your LlamaIndex RAG Pipeline with UpTrain Evaluations — LlamaIndex, Data Framework for LLM Applications

    Supercharge your LlamaIndex RAG Pipeline with UpTrain Evaluations — LlamaIndex, Data Framework for LLM Applications

    llamaindex.ai

Similar pages