Hacker News new | past | comments | ask | show | jobs | submit | halyax7's comments login

even if it was made up, its still a serious issue

I thought this looked familiar - its a reskin of a fantastic full stack tutorial from a few months back:

https://youtu.be/RkwbGuL-dzo?si=mWnIfWnmFLKIwIaD


It's not a reskin , I developed it on my own. It's just that these workflow builders are built using libraries such as React Flows. As you can see in the bottom right of the sheet of the flow builder. Also my application is built on Svelte, and I'm using Svelte Flow which is a fork of React Flow itself.


IMHO, if someone made something while following along with a tutorial, I do think they should be free to turn THAT software they made (with a guide) into a product. The presumption being that the tutorial isn't just handing you a product that you sell with no changes. (I do see there's access to the source code in the description.)

Skimming the video, and this site, I'm not noticing a ton of differences other than some minor styling changes. Maybe the author could mention any of the changes they've made? If they're significant, IMO, it's fine.


Yikes. I was wondering where all of these exact replicas were coming from. Thanks for sharing.


an issue I've seen in several RAG implementations is assuming that the target documents, however cleverly they're chunked, will be good search keys for incoming queries. Unless your incoming search text looks semantically like the documents you're searching over (not the case in general), you'll get bad hits. On a recent project, we saw a big improvement in retrieval relevance when we separated the search keys from the returned values (chunked documents), and we used an LM to generate appropriate keys which were then embedded. Appropriate in this case means "sentences like what the user might input if theyre expecting this chunk back"


Interesting! So you basically got a LM to rephrase the search phrase/keys into the style of the target documents, then used that in the RAG pipeline? Did you do an initial search first to limit the documents?


IIUC they're doing some sort of "q/a" for each chunk from documents, where they ask an LLM to "play the user role and ask a question that would be answered by this chunk". They then embed those questions, and match live user queries with those questions first, then maybe re-rank on the document chunks retrieved.


Much of those editing steps could be streamlined and/or straight up automated so that estimate will come way down over time


Consider applying for YC's Summer 2025 batch! Applications are open till May 13

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: