Grafity is a graph-based and LLM-augmented knowledge canvas that allows non-linear yet organized ideation.
Beyond basic components — nodes and edges — it also features:
- Depth-First and Breadth-First views:
- By clicking "Depth-Frist" and selecting a node, all nodes except that node and its parents will be set semi-trasparent.
- This feature is to allow users to explore their thoughts in a linear manner, while still maintaining the non-linear nature of the graph.
- AI Autocomplete:
- By clicking "AI Autocomplete", an LLM will suggest a new node with relevant label and content to connect to the selected node, based on its content.
- This feature augments the canvas in both depth and breadth. It helps the user to dive deeper into one specific node, while also reminding them of new ideas that connect.
Think of convergence with cognitive mechanisms of:
- memory facilitation, e.g. chunking;
- streams of thought, i.e. convergent / divergent;
- representation of thoughts, i.e. symbols, analogies — think of the "illogical tunnels" that seem unsystematic but actually make threads of information more intuitive.
Use LLMs built-in to understand & predict connections.
B. What input, say, what kind of data is the product designed to handle, and why is the user needing this?
Examples could be:
- handling research papers for conducting literature review,
- handling ideas for product design,
- handling tasks for progress-tracking and management,
- handling class notes & example practices for studying purposes,
- helping navigate multi-party chats (e.g. reaching out to prof, advisor, etc. simultaneously to seek for research opportunities),
The thing is, what is a logical system, or say, umbrella of system, under which all these are doable? Maybe leave this to be user-defined.
Can be extracted to image/pdf.
I.e. just like tree/graph traversal, some thinkers with explorer mindsets are intrinsically breadth-first, while others prefer depth-first thinking.
In this regard, how can the product take this into account? Should it let the user choose which mode to go with, or should it adaptively optimize representation? If we go with the latter, how?
TBD
Edges are named/classified in accordance to support Focus/Mining View in Key Fetures Section II.C
Password, Username, DisplayName, Chat (Date, title, tag, type)
User select a node, click the button, and an LLM suggests a node. This solves the "Illogical Tunnel" problem.
- Exploration View: For BFS-comfortable users. They see the whole graph, zoom in/out, and organize spatially.
- Mining/Focus View: They only see the current node and its direct children/parents. This simulates a linear stream of thought within a non-linear graph.
- The "Export map to..." allows the user to download their map/thoughts in image format.
- The "Extract Summary" allows the user to get an AI-generated summary of their current map. This summary can also be stored in and later edited in, and exported again from the "Note" space for each single chat.