If you are not too familiar with "semantic search," think of it as RAG.
To enable semantic search for text, image, etc, we need to pre-process our data before sending it to the database.
My framework works as a middle layer between your app and existing databases, where all the processings happen.
The key differentiator is its MongoDB-compatible NoSQL interface: you interact with this RAG engine exactly as you would with MongoDB, while benefiting from powerful vector search and document-augmentation capabilities.
It's open source: https://github.com/onenodehq/onenode
If there is a pocket-sized compact hardware that hosts large-size open-source LLMs that you can connect offline, wouldn't it be helpful?
The benefits:
- You can use large-size open-source LLMs without using up your PC or smartphone's compute
- You can protect your privacy
- You can use high performance LLM offline
I know there are many attempts to achieve this, but I don't know any one that actually works.
If you know any, please comment below! Thanks!
Traditionally, text and other media files such as images, video, and audio are treated separately requiring developers to set up pipeline to deal with multi modalities in their app.
Moreover, when they need semantic search capabilities over their multimodal data, they need to set up embedding pipeline with vector db.
I want to make an abstracted document DB that natively accepts multiple modalities inside JSON documents and automatic semantic search indexing.
Please check out the docs and let me know your thoughts.
https://docs.capydb.com
So... for an AI agent to be truly self-evolving, it has to have access to modify ITSELF, not only the outside world that it interacts with. This means that it has to be able to modify its source code by itself.
To do this, the most straightforward way is to give the AI a whole server to run itself, with the ability to scan its source code, modify it, and reboot the server to kind of "update" its version. If things go well, this shall show us something interesting.
https://www.capydb.com
If you don't buy upvotes, you'll be wiped out from the top list and won't even be viewed by users.
You can search and find a website where you can buy 50 upvotes for $399.
Setting ethical considerations aside for the moment, I think it can be interesting to see what happens if we have a whole server with a self-evolving AI agent inside with full permission to modify its server. To prevent shutting down itself when installing new dependencies, it can create a copy of its server and after successful installation, it can safely terminate the older version of its server.
Possibly it could generate revenue to pay its server fee and database fee etc. (Very optimistically tho)
What do you think?
This is one of the main reasons why people are complaining about the complexity of LangChain despite its intentions of making development easier.
What we end up developing internally for our own front end is a more database-oriented solution. We basically implemented the embedding and indexing functionality directory on top of MongoDB and Pinecone, making it look like a MongoDB with a built-in async embedding pipeline. The reason why we chose to do this is because we wanted to keep database-like syntax that is as powerful as LangChian.
What do you guys think? And what is your solution for your apps?
What do you think about "AI-Computer Interface"?
Let me explain a bit. "Human-computer interface (HCI)", was first developed around WWII. Since then it evolved so much and today we have very sophisticated and intuitive interfaces.
It's the AI era, and there are so many projects and products trying to automate various tasks that humans are doing. However, main approach is to try to adapt their LLM-based products to current HCI, which is developed for humans, not LLMs. This approach actually works with some workaround, such as taking screenshots to understand what actions are available, etc.
Example project: https://github.com/OpenInterpreter/open-interpreter
I wonder if creating UI specifically for LLMs/AIs instead of adapting LLMs into human-friendly UI can make it possible to achieve more efficient automation processes.
Do you know any of such projects or products?
Happy holidays!
Frameworks like LangChain and LlamaIndex offered inspiration, but their documentation often felt overwhelming due to multiple tools and versions, which added complexity instead of clarity.
After experimenting, I settled on a hybrid approach: using MongoDB as the primary database to manage RAG resources and Pinecone as a plugin for vector search. This combination offered the flexibility and performance I needed for smooth integration.
To help others facing similar challenges, I’ve made this setup available to developers. I’d love to hear your thoughts!
Documentation: https://docs.onenode.ai