Unify (YC W23) dynamically routes your prompts to the best LLM and provider, so you can easily balance cost, latency, and output quality. Just tune these three dials, and let Unify handle the rest. There is a new LLM emerging almost every week, and this means regularly testing each model against your requirements, juggling multiple accounts and API keys, and constantly updating your application to leverage the models most suited for your task. It can be pretty overwhelming. Many people give up and just try the largest models for everything. However, you really don't need GPT-4o or Opus to summarize simple documents, for example. Llama 8B is more than capable and is ~10 times faster and ~100 times cheaper. Most LLM apps are much slower and more expensive than they need to be. Founded by Daniel Lenton, Unify solves this problem by automatically routing each prompt to the best model based on your own preferences for quality, speed, and cost. Your "easy" prompts will go to the fastest and cheapest models, and only the "hard" prompts will go to the heavy lifters like GPT4. You focus on building your LLM application, and it'll focus on providing the best models, with the fastest providers, at the lowest cost. Congrats to the team on the launch!
ai for the ai
Sounds great - this will help users with efficiency and the big players with using their resources better!
The only problem I see with this Daniel Lenton is that in the future the best models will be optimized based on the input and will be dynamically priced. So most of the problem will be abstracted away. At that point you have to somehow screen for quality based on the prompt and route to the best LLM. In fact I could see there being bidding among AI models in the backend for the opportunity to answer the prompt once AI is commoditized. If you can be that middleware for the AI bidding wars, I think you’ve got a good long term model
It's looks really intresting, would you mind giving some insights on my chat bot https://www.linkedin.com/posts/sameer-m-b73376167_ai-machinelearning-langchain-activity-7198736867863724032-SWBg?utm_source=share&utm_medium=member_android
Ecosystem optimization services already. Well done on being an early mover to support adoption and choice of LLM services. It seems connected to practicality and value for the transaction - which is wonderful in these early throws of learning and iterating on use cases.
A long time ago YC funded companies like Reddit, AirBnb and DoorDash which disrupted the traditional way their businesses worked (a disruption necessary in the new internet era). Now it funds companies which route prompts to LLMs. The golden age of tech is truly over!
the API economy at work
Hello YC i have a social media idea in my mind. I have applied in YC for twice. Please review my application and give me some grants. My name is Babu Ruidas.
Co-Founder, Innerverse AI | McKinsey Alum | Google for Startups | VentureBeat Top Woman in AI
2moInnerverse uses an orchestrator with sub-agents, so multiple models from different services. A few questions: 1) How is an easy prompt defined? Does the user define this, or does Unify have a way of testing this? We need very nuanced interpretation of our data, and only certain models can provide this. It’s possible I could get the same output with a certain data feed and system prompt, but that would require a lot of testing, because we’d be putting pressure on the native NLU. 2) Our system prompt is informed by a corpus of data. Can you use Unify after an initial prompt? Our agents are conversational, so it’s multi-turn