Hacker Newsnew | past | comments | ask | show | jobs | submit | andrewlu0's commentslogin

ideally we want people to rate based on quality - but i imagine some of the results are biased rn based on loading time


That's an easy fix if you wait for the slowest one and pop them both in at the same time, no?


Nice, we're on Pinecone so maybe we can use this in the future. Could you see any way of integrating this with a token manager like Nango so our end users could integrate their own accounts?


Yeah for sure! I'm guessing you want a Plaid-like flow where end users OAuth and enter in which folder/etc?


Thanks for the feedback Sébastien! 1. Are you adding documents through the API with URLs? URL parsing is only supported through the UI right now, through the API it will just add that url as the metadata. 2. Will add md support soon! 3/4. We are working on ways to give more configuration to how the context is formatted like you mentioned, keep an eye out for updates! 5. The response from the completion API should return a "search_results" array which includes the chunks and other columns/metadata/images. If the array is empty there may be something wrong in the search config, happy to hop on a call some time or help in our Discord

Hope this helps!


Thanks, I appreciate the thorough response!

I would love it if we get more customization on the context. I just tested on a use case that is close to what our app needs for the back-end, and I'm getting chunks that are too short. It seems I can set how many chunks come up in the context, but not their length.

Other than that, you've got a subscriber. Congrats! I've just entered our CC number.


Glad to hear that! Btw you should be able to configure the chunk size and overlap when you upload through the UI, or when using our /upload API (https://docs.baseplate.ai/api-reference/documents/upload-doc...).


Perfect! I played around with this and got exactly what I needed. Thanks :)


Yea! Our FE is React/Next and BE is a mix of Supabase, Pinecone, BigQuery and QStash. and mix of HF endpoints and PoplarML in our batch for some embedding models


thanks for using Supabase andrew.

Are you using it to store the embeddings (pgvector)? I'm asking to see where you're differentiating between supabase & pinecone. We get this question often and it's interesting to hear what others are doing and where they feel pgvector isn't appropriate


Hey! Loving the experience so far - we started the project before pgvector was supported on Supabase, and we want to support hybrid search as well (don't think pgvector supports this?)


> hybrid search

Yes, I believe it can be done but it requires a bit of work on your side. I'll see if we can come up with a demo and/or postgres extension which handles this for you


looping back here. The langchain community just released a version of Hybrid Search which using postgres full text search. It's clever, and IMO probably a better approach than just sparse/dense vectors:

https://js.langchain.com/docs/modules/indexes/retrievers/sup...


Will check this out - it seems like its doing two separate searches where you specify two separate top K values. Curious what the trade offs are between this approach and weighting sparse/dense vectors


Thanks guys for sharing it. This really helps us in the community to look for the architectures that we should focus on and the new interesting developments.


Right now typical usage is < 100 MB per org, but theoretically it could infinitely scale with auto-scaling. And yes heres a new link: https://discord.gg/JCqEzCZ4FM, should be updated on our site as well


It should let you trial without a credit card - let me know if thats not working!


We think our current product makes it alot easier to work with documents/embeddings to help build that initial prototype, and once its deployed, tools for versioning and logging. One example use case is ingesting previous support chats and docs to build a "copilot" for your support team.


Yes, you'll have to bring your own API keys


Hi, when you create an endpoint it should give you API examples based on the deployed app. Also docs here: https://docs.baseplate.ai/api-reference/completions. Hope this helps!


For standard datasets we use OpenAI ada embeddings, for hybrid its instructor + SPLADE. The HyDE toggle in the context variable feeds the query to a prompt first ("Generate a document that answers..."), before embedding. I think in the paper they use the contriever embedding model but we just use the ones supported on our platform


I think we will be hearing a lot more about HyDE. It’s a neat trick.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: