
Tech • IA • Crypto
Graphify, un plugin d’IA en forte croissance, utilise des graphes de connaissances pour créer une mémoire persistante pour le code et les données, réduisant l’usage de tokens et améliorant l’efficacité pour les développeurs et les entreprises.
Graphify a dépassé 500 000 téléchargements en quelques semaines et atteint environ 43 000 étoiles sur GitHub, signalant une forte demande pour des solutions de mémoire IA. L’outil est apparu rapidement après une idée inspirée par Andrej Karpathy, avec une première version construite en seulement deux jours. Son modèle open source a accéléré l’adoption communautaire et l’ajout de fonctionnalités.
Le système transforme bases de code, documents et contenus multimédias en graphes de connaissances structurés, reliant les relations entre les données. Cela crée un « cerveau numérique » persistant qui conserve le contexte entre les sessions, répondant à une limite majeure des grands modèles de langage: la perte de mémoire entre interactions. L’approche combine réseaux neuronaux et structures symboliques, appelée IA neuro-symbolique.
Graphify cartographie automatiquement des bases de code entières via l’analyse d’arbres syntaxiques abstraits, reliant fonctions, dépendances et fichiers dans une structure unifiée. Les ingénieurs peuvent ainsi comprendre des systèmes complexes en minutes plutôt qu’en jours. Pour des entreprises payant des développeurs seniors environ 150 $/heure, cela se traduit par des économies significatives et un onboarding plus rapide.
La structure en graphe permet de tracer les dépendances et d’identifier rapidement les problèmes. Au lieu d’enquêter manuellement pendant plusieurs jours, les ingénieurs peuvent repérer instantanément les relations entre composants. Cela réduit le temps de résolution des problèmes backend et améliore l’efficacité globale du développement.
Au-delà du code, Graphify traite documents, audio, vidéo et URL, les convertissant en graphes consultables. Par exemple, des enregistrements de cours peuvent être transcrits, segmentés et cartographiés, permettant d’extraire des informations précises sans revoir l’intégralité du contenu. Cela crée une couche de mémoire unifiée pour des données variées.
L’outil limite les requêtes répétées avec contexte complet, réduisant fortement l’usage de tokens. Les gains varient selon l’usage, de 20x à plus de 90x, avec 70x dans certains cas. Les performances dépendent largement de la structuration des requêtes et de l’interaction avec le graphe.
De nouvelles mises à jour prennent en charge des backends locaux via des outils comme Ollama, permettant d’exécuter des petits modèles de langage (SLM) en local. Cela réduit les coûts et améliore la confidentialité des données en évitant le cloud. Les utilisateurs peuvent choisir entre modèles locaux et cloud selon les tâches.
Graphify vise à créer un jumeau numérique des connaissances organisationnelles, capturant discussions internes, documents et flux de travail. Des fonctionnalités en cours incluent des intégrations avec Google Workspace, AWS, Slack et les transcriptions de réunions, pour maintenir un graphe de connaissances constamment à jour.
La plateforme intègre des mises à jour incrémentales du graphe et des mécanismes de déduplication, ajoutant efficacement de nouvelles données sans retraiter l’ensemble. Cela permet de monter en charge avec l’information tout en conservant performance et précision.
Les recherches en cours incluent des embeddings hyperboliques pour améliorer la récupération hiérarchique et réduire la perte de contexte dans des graphes profonds. L’objectif est une navigation plus précise dans des structures complexes sans dégrader la qualité du contexte.
Graphify illustre un basculement vers des systèmes de mémoire IA structurée, combinant graphes et modèles de langage pour améliorer l’efficacité, réduire les coûts et instaurer une connaissance persistante.
If you are using Claude Code, you've definitely heard about the new plug-in called Graphify, which is essentially a memory solution for all AI users. I already did a full breakdown on Graphify itself, so if you haven't seen that yet, please watch that here first. But this plugin already has over half a million downloads in just a few weeks that it was set up. And boy, do I have a treat for you guys today. You're going to have to excuse some of the audio mixups, but I have the creator of this plugin here live with me. And together on this video, I'm going to be asking him a lot of the same questions that you guys have in mind about this plugin. It's going to be interview style. We'll go through his best use cases, his inspiration on why he created this plug-in, and pinpoint the truth of whether this plugin can actually save you 70 times on token. So, without further ado, grab your cup of Joe, grab a snack, and let's get into this. All right, guys. I got Safi here with me. He is the creator of Graphify. And today we're going to go through some common questions, kind of reveal the intentions and inspiration behind creating this awesome plug-in. So Sophie, what I'd love for you to give us is an elevator pitch. I got some questions on my side, so you guys might see me looking over. Uh, but give us an elevator pitch, a TLDDR of what you currently do and how that work really led to the creation of Graphify. >> Hi guys, Sophie here, the creator of Graphify. So I'm I'm an AI engineer been working on knowledge graphs since over two years now. I've got two research papers on graph networks and then hybrid rock systems have combined vector databases with graph architectures and I've been working on the increasing the process memory systems of large language models and that's where the graphify came in as a link to my research and then my AI engineering experience. Yeah. So graphify actually started as a link by from a post of Andre Karpati. Uh so he tweeted about like a structure when you can use raw data files and then you can extract them later on so that you can minimize the tokens usage of your uh coding assistants and that's where the idea the whole idea clicked in from my experience of have been been working on graphs. I shipped that feature within two days of Karpati's uh uh idea pitch on to X and that's how it all started and here we are with 500K plus downloads on Pipi and nearly 43 43K stars on GitHub. >> I absolutely wild how that actually all went down. Um what I'm trying to understand now right is like did you did you expect the kind of traction that you got when you guys when you initially launched this product? Was it kind of expected or were you really surprised? >> Oh, I'm totally humbled by this totally. So I I did think like it will go past 10k stars maybe because it was a need as a tool it was in need of an hour and the the timing was to totally correct and then right after fe of kapati I did expect like it should it will blow up a bit but not this far and and I'm totally humbled by this uh the the people the sgestions of people and then people have been using this and putting it publicly as well how they are using my my my software graphify they've been making posts on LinkedIn they've been making posted eggs and and pretty much uh what do we say I'm I'm in dilemma of like it's it's something I had never experienced before and I'm pretty much excited about this part. >> Yeah, that's key. You know, I I think before honor comes humility and keeping the humility is always key. It's what makes us likable and from what I've known you I've known you that this past month, you are a super humble guy and I think you deserve everything that you have coming for you. and looking at what we're looking at what for sure looking at where things have come >> like walk us through well not this part yet what I want to ask you is when you were looking at this was this you per like were you scratching a personal itch when you kind of put this together um was this was this aimed at a specific category do you did you have something in mind when creating this or was it more like you saw a problem you solved it and you were kind of like testing your skills Tell me a little bit about how that looked. >> So it all started as a personal as you told as I told you I'm an AI engineer. I've been working in this in this whole system since over two years now. So I've been facing the same problem of like losing the context once I switched from one session to another session and and and AI labs like anthropic and open are continuously working on fixing this problem of context when you switch from one session to another session and they've been giving a compact conversation feature as well. you must have seen and and the whole reason I came up with this idea was like let's do something into it. let's come up with a good good feature good good software that can fix this problem for just not me but for others as well that's necessary and I came up with the open source software idea to do this and then eventually when I saw people picking up my tool I I went full on it and then started shipping as fast as possible started integrating more features did some R&D and then went into production after that it's it's doing pretty well at at this point uh it's been like a month now and I've got like decent decent uh response from the audience Yeah, you definitely have gotten a really good um amount of traction on this. So, now that I have a little bit more of an understanding, which I get because that's usually where all things start and the the market demand was there. You know, we all knew it. Memory is a huge problem. It's what a lot of the AI users are looking at solving right now. Nobody has quite solved it yet. I wanted to segue into something that I think a lot of us are really curious about because there's beginners, there's intermediate, you know, uh, AI users, and then there's more advanced. And I want to know, you know, what are the what are the if you had to pinpoint the three best use cases of this product, maybe for a business owner, you an everyday college student, you know, like different categories of people, like what would you say are the top three use cases for Graphify in your head? >> Yes. So, let's say Graphifi's main criteria is it's it works pretty well with your coding assistance. Let's say you've got a new uh contractor employee who's going to come as a senior engineer in your in your company and then he wants to go through the whole codebase of your of your of the whole architecture of your company's codebase. So if he were to get in deep into the whole architecture it will take him like two days three days and plenty of slack messages to the previous employer or something like that. So just to prevent this whole scenario to come like to spark this whole scenario graphify it what it does is it creates a map of the whole codebase using abstract syntax tree a libraries it's it's free of cost as well there is no lms involved into it so it creates a map it creates god nodes and then it connects with the coh different parts of a codebase together and then you can pretty much pretty much easily understand them in one shot it saves you time saves you money. Let's say you you you are hiring someone for like $150 per hour just to go through your code base and you can ship that you can you can do that for free in like few minutes and it's a big ROI just by a tool. Secondly, uh let's say you are working for an enterprise and then there's a ticket on your back end some some users facing a problem right and then you want to fix that ticket as soon as possible. What are you going to do? uh maybe one of your engineers will sit or fix this for the whole day or two days or come up with a solution after 3 days right so to to like bring the time down here you all you have to do is you have got a brain here you got the map of the whole code piece you'll just look for the dependency and you will see how the dependency is connected to all other files in the same codebase and and it's one shot it's boom you've got the whole solution from there you can easily find the solution from the god nodes of the whole architecture thirdly uh so graphify isn't just about the code bases It's also there to create persistent memory. So we create a map of your documents, of your images, on of your YouTube urls, of your audios using whisper. So I'm converting all those videos and audios into transcripts and then creating those text from those transcripts in the form of a graph and you can pretty much retrieve everything. Let's say you had some lecture or you want to see you you want to prompt a lecture of of of Stanford let's say and then what you're going to do is just put that link get the get the lectures uh transcription and then you can have the map of the of the lecture with each section of the lecture divided into into pre various graphs and you can retrieve anything much from there. It saves you it saves you a lot of time like go through the whole video rather than like prompt the whole video in one shot. So it saves you time and money both of these things and gives you persistent memory as well. You can keep on adding context to the same graph time after time with less tokens rather than like making the coding agent go through the whole corpus uh or call multiple times for a simple task then going at once and then shipping you the final response. >> Right on. And what what I'm looking at is for me what's really really exciting is the ability to create the semantic relationships meaning the hidden meaning between the different assets that is super huge because when I have something searched in my clo my claude files it's like how does this file really connect to this file you know claude can search and find the different file names but I think a really good important use case especially on number three which is for everyday day-to-day users users. That relationshipbased mapping is absolutely huge. Now, what we're looking at is essentially a local rag system. You know, it it's a database. It's a way to look through all that information. And now with the development of of all these different plugins like GSD you know about you know you might have seen some of the stuff we've built internally on our side Paul and seed you have non-engineers non-developers building develop you know actually developing products so this is also a way for people to use these these plugins and these orchestration layers to build products match it with something that allows them to sift through their code bases more effectively with less tokens and and to have those relationships tied together creating a a way more effective and I even fun like I mean using these tools are so much fun >> just in exponentially increasing our output and our our capacities capabilities so on that note let me pull up my questions again um so we understand >> I have something to say here uh that's the best best use case here. So there's too much of AI slop coming in. People are shipping fast using AI and they are actually aware what what's what they are shipping the code. There's a lot of junk code there. So what if you like map the junk code and then you can you can see what's what's wrong in your code as well by just using a small coding tool coding assistant graphify. So it can help the non-coders with no technical background as well to like retrieve and then fix their fix the issues with their code bases. And that's one of the use cases I I forgot earlier on. >> Yeah, there's definitely a lot of AI slop out there, bro. Um, so it it that question kind of tied into my next question, so I might not have to go deep into it, but maybe there's a uh specific ROI that you can outline. you know, for most business owners, most of my clients, you know, these are guys that are doing at least a quarter million plus a year in their business and they want to know, you know, because we've put out Charlie OS and that's, you know, oneclick install for claw code. You know, the frameworks, the file structure, you know, how important file structure is using clawed code internally. If that's not set right, like everything else is just going to be a mess. You you you need the right frameworks. So for those of for those of for those business owners that might be watching this or anybody looking to start a business um is there anything that you might have missed in terms of graphify like how how is this save we know how it's saving time but is there anything that you might have missed in what you said about the particular ROI financially for businesses in this case is there anything you want to expound on there >> yes so as an enterprise let's say uh you have hired a new person and then uh so basically The whole concept is about a digital brain. So we create a digital twin of your whole enterprise. Let's say uh see the context uh doesn't last for long. Either it says in your head or it says on Slack messages. It is people forget forget what they talked about yesterday. So graphify has a capability of giving you a persistent digital twin of your whole corpus which you can retrieve at any certain point of time rather than wasting your time and money hiring someone to search things for you or do things for you uh which you can do just by using your own coding assistance. So you're going to save a lot of like uh unnecessary money having been spent on employees doing very trivial task for you where you have your own digital twin your whole enterprise brain working also graphify has an incremental update feature which updates the whole graph automatically with every new injection of documents you have so it's an intelligent enterprise brain which gets smarter with time so I'm also working on a new layer of graphify which will make the whole brain smarter with self-arning layer I'm going to implement there so graphifi's brain will adapt to the users's use case like a certain domain let's say you are working in legal tech or rather you're working in real estate so the whole graph will start adapting as per your own requirements and that's my uh coming vision future vision is to help the enterprises as on bigger levels >> yeah I I love how excited you are about this product because a lot of the guys that I've been meeting on this journey, uh, posting on social media, it's only been about 5 months for me, but meeting you and seeing this product really, really push far and especially the one video I did, it has a little under half a million views. That was a huge accomplishment for me. Um, so thanks to you and your product, you know, I was able to accomplish that. So, you know, it's really exciting hearing about the new developments. I know you're very excited about the enterprise solution and I I think with the right push and uh hopefully guys if you guys are watching this and this gets enough traffic definitely like the video. If we get over a thousand likes on this video, he hasn't promised to do another video with me yet, but if this gets over a,000 likes, he might be tempted to do another one with me so we can go deeper on the workflows and see what he's got going on. Um, so anyways, the first version that dropped, right, and then we're going to go into something a lot of you guys are probably curious about, like how would a new user use this if you're starting from scratch? We'll get into that. Uh, but I want to ask you with the version one that dropped versus what you have now, what are some of the key differences that you've actually put into the news version that make this even better than when it first launched? >> Yeah, so there there are plenty of differences now. So I have given access to local backend system as well. So you can use Olama back end as well. Now earlier on it was just like you you were using CL cloud-based LMS which were given by your coding assistance like GPD or so models or like Opus models. So they were like causing a lot of problems to users on the burns of to burn of tokens. However, now you can call on your back end using llama uh local model as well the SLMs and it's pretty much cost effective and I've also got like 27 plus languages now earlier on we started with like 50 plus languages so we have gone past 26 plus languages or for the as you can extract a lot of languages uh coding languages now uh so backend system with uh a local local SLMs or LLMs and then uh with with more coding languages thirdly I have released another feature You can connect your Google workspace or even your AWS bedrock models there with graphify as well. Now it's it's a big solution for the enterprises they were looking for. So you can even collect bring your documents from Google workspace directly into graphify create a map of those documents and then you can retrieve whatever you want. See the relationships on your offse as well if you want. It's more interactive. >> That's crazy. >> And visually appeal. >> Yep. I I I'm using GWS CLI. It used a term called SLM. I'll have to research that. I know what LLM is. What is that? Small language model. >> Yes. All right. That makes sense. So basically large language models do not fit in smaller RAM size. So you need larger RAM or so GPUs like CUDA kernels to fit them in. So rather than like going for LMS, you can just go for SLMs. And SLMs are the future by the way because everything is going to be local and and that's what people are keen on. so that their data isn't shared to cloud-based LLMs in the future. And that's my vision is at the moment. >> Yeah. The the the good news is that the uh the small language model, the local stuff is not any worse. >> Yeah, it's a local first AI memory system. That's what Graphy will be in coming days or months. Pretty much none of your file will be shared with cloud-based LMS. It'll all run locally on your system. And that's the best uh win for the enterprises in the coming days. Yeah, that that's exactly what I'm what I'm seeing is when we get to a point where the large language models become they're going to keep getting better, right? Always better and better. But what's going to happen as a byproduct is that these small local models, this is the worst they'll ever be and they're only going to get better. So, a year from now, a small language model locally hosted might be as good as Opus 4.6 or maybe two years. I don't I don't know. What would you say like when are we going to have a local model as powerful as let's say Opus 4.6? six or 4.7 like what is your take on that >> perhaps maybe a year or so with the pace we are moving forward I think so we are going to have pretty smart SLMs pretty soon so basically what I've done is I I've fine- tuned SLM for my use case and they work decently well as as large language models as well so it's like how could you fine tune your SLMs for your use case >> so what I want to ask you now is I want to look at what do we uh what would be the best way for a new user maybe who's never used this before to effectively set this up, you know, because I know the biggest for me was like I set it up and it ran through my whole daily limit and then I I spent like an extra like 25 or like 30 bucks off of and I'm on the $100 monthly plan with with Claude. Um what would be the best way to effectively run this without killing your whole limit and just exceeding it? Like what would you recommend? would you say um one of the gentlemen I did an onboarding call with before this he's like do I uh do I pretty much like say x amount of files do create a handoff do the rest of them do I just power bang you know my whole database um and am I even ready to run graphify for a new user it's like is it is it effective for them to do it with nothing in their codebase especially if they're not like developing things that's uh that's a lot I just threw at you um but does That question makes sense. >> The main problem if you are running if you are calling an LLM to parse your code base that doesn't make sense at all. I've given I' we have got like a there. So so you don't have to like call a cloud-based LM rather than like it's free of cost. The a call is free of cost. There's no API call included for an LLM. So a pretty work pretty much well for the code bases. You have to call the LLM back end only for your documents. Abstract syntax trees. It's a free Python library which can connect various parts of a codebase like functions and all of those sections of a code all together and create uh relationships between them in different files even in multiple repository. So it's it's free of cost. So basically what you have to do is with the new feature I've added in graphy with backend system is Olama. So you've got Lama as well as a back end which is free of cost as well. you've got locally running LLMs or SLMs for you or using Olama's API. Uh so what you have to do is call the code base using as which which which works pretty well for my even doers as at the moment and then if you have some documents you can use an LLM. So which creates a semantic extraction of the corpus of the of the chunks within a document and that's how you can save a lot of money. Secondly, when you are querying that the the graph the map, make sure that you first ask you you prompt to give you the god nodes from the whole corpus first of all. So so if you have large amount of god nodes that means that something has gone wrong with the cohesion or or maybe with your codebase. Uh if you are looking for an extraction of your codebase as well. uh and if you have a a decent number of god notes maybe for documents as well things have worked fine and then you have the whole architecture ready right in front of your eyes of even of your documents as well of your codebase just by prompting for god nodes and secondly from god nodes you can easily see where you have to prompt and write to minimize as much tokens as possible uh third also if when you are like ingesting more data make sure you use the word uh update as well graphy update so that it doesn't like start fresh from the beginning rather than I've used hashing strategy h 256 in here so it will start from where it left and then I've also got some dduplication techniques so there will be very less collusions of the entities within the same document or multiple documents so I've taken everything uh taken care of everything all you have to do is follow the right tactics for the right hooks I have put in the readme file or maybe you can just ask cla code to explain how to make the best use of graphy uh as the initial shot and then you work from there as well as a as a work tree workflow for for your whole use cases. >> Yeah, that makes sense. >> Yeah. So, it's a always remember as for code bases, LLM's for documents, extract backend using llama or if if you want to save cost or if you want to go with all your cloud-based LLMs, you can choose cloud code, gemini, openis, any of the models GPD. Uh and then after that always go for the god nodes and from there you can work work through into anything. Also make sure that you don't have to like when you are prompting the graph don't ask the LLM to go through the graph rather ask the LM to extract from the graph. So graph is the memory graph is the context. So always make sure and you can work around a bit and I've seen people so there is no ceiling or floor in token uh savings here. What I've got 71.5 in my read me is for the repository I tried on and I've totally mentioned that it I've seen people who are getting like 90x token reduction in their retrieval as well and there are people who are getting 20x token reduction in their retrieval. So it's totally corpus dependent and and there is no ceiling or floor here. >> You know what's interesting? I have a lot of hate in my comments and I'm sure you've seen it. Oh 70 times what are you out of your mind? I'm like, "Guy, relax." Like, all these all these different situations are circumstantial. It's like from what you're saying, right? >> You and let me let me just let me can I I'm going to try to put this in layman terms because, you know, you're you're a super smart guy. For me, a lot of what you're saying, I'm reading in between the lines. Like, I hear big words and I chalk it up to like, okay, that he's just saying like as for code. So if if uh GSD builds an app, okay, that's perfect for that. So in a n like if you had to explain it in like an an to a ninth grader, you know, as having it look through codebase, great. Using um you said use an for your documents >> and you mentioned >> uh >> llama back for documents which are also free of cost now. So it it gives you a local model running on your own RAM on your own whole machine rather than calling a cloud-based LM. >> And I and I know that a lot of this is already hooked into Graphify. So if you guys have any questions on this, you guys can drop a comment below. And what I would suggest to get the best use or to get the best out of this, take the transcript of the video. What are you waiting for? plug it into your clawed code and when you guys are setting up or using Graphify, this is going to be perfect for you guys and um I think this is going to be great for anybody looking to use Graphify. So, on that note, I I I want I don't want to take up too much more of your time. I know you're excited about the growth. You know, I can go on to the the marketing side. I don't think you need help with the marketing right now. That's it's exploding on its own with the uh the creator outreach. So, we talked a little bit, maybe we didn't, about the uh the future of AI memory. So, there's two more questions I want to ask you. And then there were some some of my actual followers and school users that had particular questions and that isn't actually super long. Um, but before I get into these two questions, I just want to expound on what he said because he said something super important here. Sophie Safi Sophie was saying that each of these situations when you're using Graphify, it's circumstantial. There's no floor. There's no ceiling. So now when you guys are using this tool, it's dependent on how you're leveraging it. >> That's how you're going to start seeing the savings. So if you're not following this this playbook here, you know, he's talking about god nodes. It's like highlevel uh nodes, the god nodes, right? If you guys aren't leveraging that and knowing how to input that when using claude, it might be running and looking into all the specific nodes. So if you want highle information, you can then scalpel like go into the graph and get the information you need with those semantic relationships tied in. So the proper understanding and context and what he's saying of 70 times usage is key when understanding whether that's actually valid or not. So, I just want to leave that there and I want to tie into the next two questions here before we go into the the user questions. Where do we see the future of AI memory going? And I just want to hear your thoughts there and I want to understand after you share that, what is the biggest takeaway you're trying to give a user who's watching this video? what is the biggest takeaway you can give them that'll 20 times output their usage with with their their language models maybe let's just say cloud code so two questions there >> yep so the current future uh with AI memory uh always be a problem to be honest if you have worked with the whole structure of how neural networks are set and how retrieval systems work so basically there are some people some startups who are working on the same problem like super memory they are based in SF hydrab And then there's another thing called Gbrain by Gary Tan who is also working on the same problem of large context for LLMs and and and it's the market is very bright in in in person memory systems and I believe there's too much to come in next three years and uh the problem will never be solved until unless we have a new architecture other than transformers which are the base of NLM at the moment because there always be hallucination there there will always be a loss of context and uh they but they they never will be 100% efficient as as a human and that's something uh will always exist and persist but you can always optimize with the right tools with the right techniques as I said graphy is one of those tools if you optimize it the right way you can always have a persistent memory of your code base of your documents with incremental updates and features >> okay now I hear you on that I think how we use things will row like how much MPG in a car miles per gallon that has we're never going to get unlimited miles per gallon but we can always raise that ceiling. We can change the source of the gas you know the the the power source from gas to battery. We can extend those miles with different uh infrastructure and and setups. So I see where you're going with that and I wouldn't disagree at all. So what what would be your biggest takeaway like to anybody watching this? What do you want them to know that's in your brain that would allow them to use graphify and totally just change like total game changer using graphify? What would that takeaway be? >> Takeaway would be first of all you should know what you're putting into the whole system and if you're putting codebase you should always aware that graphify has uh in there so you have to like uh spend your tokens just to fetch the the code base. So, play it smart. Uh, and then you can have a lot of savings in there. For documents, maybe use a as I said earlier in in the interview as well. You can have a SLM now working for you using Llama. Uh, and it's free of cost as well rather than like going for a cloud-based LM that takes a lot of money. Secondly, you can split your repository or your documents as well and then you can reingest them uh in a way they they will just get added in the best way possible rather than like putting the whole chunk of documents in one shot. So basically we say that claude has 1 million of context and everyone is going towards like maximum millions of context in in one session right but it doesn't make sense at all. the more the context would be there will be a massive chance of hallucination or dilution in the reasoning as well and that's where you lose and that's where the whole graphite system comes in. So graphify isn't just uh supporting neural networks it's giving a s it's giving a a rise to neurosymbolic AI systems where you have a map or symbols to support the neural networks to come up with a response. So, and that's the whole graph vision is neurosyolic AI. >> Neuro symbolic AI. I love that. I it it's a way to simplify the language of the data being processed in a way that makes it efficient and and more relatable. No, >> not really. Yeah. So, neural networks, right? And then neural networks can hallucinate as well. So all the retrieval all of those uh path pathways can can hallucinate any response. So to make this whole problem uh with solvable we have symbols. So symbols are in the form of graph structures and then if you go deep dive into this graph theory you'll know about the new symbolic AI systems and even graphs also can give you an opportunity of clustering various communities in there which other tools like obsidian would never would never able would never give you like I've seen like some people are asking can't we just like make obsidian do do this and that it's not possible at all obsidian can't do clustering for you obsidian can't do cross community interaction for it will just give you an visualization which is which looks pretty much well but there is nothing credible you can take from there so you always need you always need a new symbolic structure or a tool that can solve it for you in one shot and that's where graphy is >> yeah I love that because that was literally what what a lot of my guys are saying to me bro they're like well can I just use obsidian by itself but this is this is basically putting in a way that makes sense like when it's in graphify it's just ass is this is just what already is existing versus having it organized and sifted through with your tool. So, I love that. I think you hit the nail on the head with that. And what I want to look at, you know, we we talked briefly now about the >> uh I'm working on another research for graphify, but I'll be doing hyperbolic embeddings uh using point ball theory, which is a pretty much a famous problem in mathematics from 2002. So this will solve the problem of retrieval hierarchal retrieval as well. So basically when they are god knots on top so they are baby nodes the child nodes at the bottom and then when you when you go down into into a tree there's always a context drift. You lose the context the more deep you go into a tree. And to solve that problem, I'm working on hyperbolic mings which will give you an exponential downward flow into a graph into a tree without losing context. And I think I will be implementing this whole research uh what I'm doing with graphifi's code base as well which will help and leverage the use cases of my users currently on on the open source side as well. >> Yeah. What what we're going to have to do man is when you come to the states we'll have to meet up. Um, I'd love to do like a um even if it's maybe I'm I'm gonna be in Florida at the time. I don't know when you're coming, but if we have something upstate in New York or something in New Jersey, uh what we could do is we can plan ahead. We can invite people out. We can maybe rent an area kind of bring some awareness >> next month because I' I've won a hackathon last last week and they are taking me to SF uh maybe in June or July something. >> Okay. Yeah. Well, whenever you come, man, we we'll have to figure that out because I I definitely want to do some local meetups. Uh but to kind of wrap this thing up, I have some um some basic questions here. We talked briefly about the Obsidian overlap. Um it's not a hard dependency. We know that it it is I I would say it's a decent recommended pairing because it's visual. >> Yes. Contact people obviously I think is is is a decent way of visualizing your stuff. and then it totally makes it more interactive rather than like going through the whole just on files and then extracting stuff. >> Yeah. Yeah. I I would agree with that. Um we know that Obsidian by itself is not going to get the best out of their data. Um so I what I want to ask you is we talked about the token economics that was a big question. Um something anal ad in my school he asked why does why does graphify not support scripts >> shell scripts are very flat I think I believe and they're very indiscriminatory I would be adding those as well in the future but for now if you want you can just pass them to an LM so what LM would do is rather than like going with an A which isn't capability at the moment for shell scripts what you can do is just put those as documents and then make the LM do a semantic extraction of a shell script so it'll be able to fetch you all the nodes from there. That's a that's a fix for now, a temporary fix. I'll I'll get it sorted maybe in coming releases so you can have shell scripts as well in there. But to me, shell scripts are a bit flat and that's the reason I didn't put them in the beginning. If if it's a requirement, it's a hard requirement, I I'll for sure put that in and I'm I'm more open for PRs and issues being submitted on the GitHub repository. I I I fix them on time and I'm pretty much active at the moment. >> Bro, I know you're always moving. every time I talk to you, you're always you're always moving. Um, so we kind of we kind of hit briefly, right? Like what would you do in your first three minutes of Graphify with a fresh uh client codebase which you've never seen? Which views or curies do you hit first to decide whether the architecture is sane versus a landmine? >> Yes. So you're talking about codebase, right? >> Mhm. Let's say you got a new hire as I told you in the beginning as well who has never seen the architecture and then you have been running graphify on your own code base even of on your multiple repositories of your whole enterprise where it has been incrementally incrementally updating the nodes and even the god nodes even in the child nodes and then clustering them even merging or connecting various communities together how they are dependent on on various dependencies in the codebase. So it's pretty easy for a new hire or maybe a new contracted senior engineer who wants to go to the code base and then have a look at the at the whole arch whole band architecture in in one shot, >> right? And I I think it's pretty self-explanatory when it when it's ran. It's like you're best off just running it through and kind of just seeing the output. I mean, if you have a subscription, it's not like you're paying for API cost. So yeah, that was a great answer. Um token economics we've went through. Um, for the people who have asked why is it that they're not seeing any token savings, it's because of the way that you're using it. We we if you know we've discussed that earlier in this video. Um, so I think token economics practical first run obsidian overlap we've went through. Um, and the last question is what is your current workflow like? What is it that you do like using this tool? Like what's your day-to-day look like with graphify for yourself? So current workflow is uh as I as I said I've been using local lms since earlier days but the thing is I didn't put them in since I was doing some hardcore testing if local are going to work good in a good way or not and and I've seen after some testing it's working working pretty well. So for the documents as I mentioned earlier on as well so you you use a local lm and then you can extract the information from there. Also by the way I'm adding a connector to various uh even even your one node even your slack I'm working on connectors as well. So you can connect your slack messages and then you can have them uh as as a map as well from from now onwards in in the coming releases even your slack messages even your meetings the meeting transcripts everything will be in a brain and as and you can work as an enterprise it will be on one short retrieval. So my current workflow is as I said local LM for the extraction and then always go to the god nodes first and then start retrieving from there and make sure you pron the graph and not like the LLM to go through the whole corpus once again and that's one thing also I'm adding some more features so that the the like the the agent would be hardcoded or hard configured to go through the graph first rather than doing anything else. what if what respect to what you prompt when graphy is active in a session and that's something I'm I'm looking forward to as well. >> Yeah, I I think finding a way to properly get it to work maybe via a hook or maybe in the cloud MD those instructions with it effectively like if if there was a way to have that integrated with an optimized flow to save tokens and it worked great like that would be phenomenal. I think some of the for me at least right now it's like anything the static information the information in your codebase or for me using it as like an operation center like my clients my data like everything and anything I I do I have MCPs CLI tools but my assets become stale over time so the process of having to remember to you know groom your data and go through a lot of that does um for most people um become a little bit of a constraint. So I'd love to see you know a lot of these things that you're doing over time to solve. I know >> always one more thing just hit my brain always make sure you update you you you put graphy update as well when you are like ingesting new files or code bases or documents as well. So make sure that it gets the right injection rather than when you you ask a question it doesn't have to go through the whole code once more. So you have to like update it as well. some of the techniques. Uh maybe I can share you a file and then you can share it with your audience as well. Like the best way of using graphify would be pretty much helpful like the way I do from the creator itself. >> What I want to do and I didn't mention mention this earlier but he's dropping a book or he already has it out your new book. >> Yeah, it's out on gum road. It's called the memory layer. So it it talks about the whole journey as well how I've been working on graphs and also tells you about how graph has been in in like in computer science you can you can learn about knowledge graphs and graph theory spectral graph theory the matrices the llassian matrices and everything the community detection strategies using clustering algorithms lead in low vein so everything every theory about graphs is there and I think it's it's a it's a much must something must you can go for >> the link is going to be in the description below And if you did have an additional PDF or playbook that you put together, I can host it on my website. Um, so if you guys check the description, the book will be there. I might have a resource depending on, you know, whether he wants to drop that sauce or not. Um, but I know Sophie, Sophie, I know you're busy and you've answered so many questions for me personally. I'm going to take this transcript and throw it into my quad so that I can self-improve. What I love what you're doing is connecting the different like grid AI transcripts or you're you're pulling in Slack and you're creating you know nodes for that and you're going to create that on like an enterprise level. I think that's absolutely huge and I think there is going to be a lot of ways for us to collaborate in the future. Uh but I don't want to hold you up. So if there's anything that you want to add that you want to leave people with before we conclude this thing I I would love for you to you know share your last words again I know you're busy so floor is yours >> so to other fellow developers I just I've just got one thing to say you can just do things and if you just have like the best determination in you that's it you can just do things >> yeah we could just do things with Claude now the best way to put it so guys uh if you can please like the video. Subscribe if you like the kind of content that you're seeing here. And we're going to Sophie and I are going to try to do some more videos together if he's got time. But I'm also going to link his social medias below. We'll link Graphy. Um but again, everybody just leave a comment below. Thank him for joining us here today because he's really busy. He's doing a lot of great things in the AI memory space and you got to remember his name. You got to remember this guy. He's going to go far. So Safi, thank you for joining us, brother. Greatly appreciate you. >> Thank you so much, Charles, for having me here as well. >> All right, brother.