Until now, customers have had to choose between RDF and Labeled Property Graphs (LPGs), accepting the tradeoffs and limitations of each. RDF offers robust semantics and established standards like Web Ontology Language (OWL) and SHACL, but it can be complex to manage. On the other hand, LPGs typically run faster for large-scale analytics but may lack the rich semantics that some enterprise scenarios require.
DataWalk changes the game. It is a no-compromise platform that provides the performance, scalability, and simplicity of property graphs, the modeling flexibility of RDF, and additional breakthrough capabilities such as enterprise prototyping, dynamic inferencing, and no-code analytics. Whether you’re integrating existing ontologies or building something new, DataWalk gives you the structure you need and the tools to work with it.
This paper provides architects and other technical experts with an in-depth review and comparison of RDF, LPGs, and the unique DataWalk approach.
The table below provides a summary comparison of RDF graphs and labeled property graphs.
Criteria | RDF | LPG |
Scalability/ | Harder to scale for large analytics workloads | Performs well with big data |
Simplicity | Can be complex | Yes |
Flexibility | Easy to modify schema | Limited flexibility for complex ontologies |
Inferencing | Built-in inference and reasoning capabilities via formal semantics | Limited |
Other |
|
|
DataWalk eliminates the need for you to compromise with your graph technology selection. DataWalk gives you the performance and simplicity of a labeled property graph, the modeling flexibility typically associated with RDF, and adds additional capabilities:
With DataWalk, you get a no-compromise platform that’s fast, flexible, and built to adapt.
Figure 1: DataWalk Provides The Best of LPGs and RDF.
In most LPG systems, you can attach properties to relationships, but you’re then locked into your initial modeling decisions. Changing the structure later often means updating the schema and rewriting parts of the graph.
RDF offers more flexibility but brings added complexity and performance overhead.
DataWalk solves this differently. With DataWalk’s Connecting Sets capability, you can evolve your model over time, without needing to rebuild it. The example below illustrates this.
Step 1: Modeling Responsibilities Between Parties and Accounts
Let’s assume you need to model which parties (people or organizations) are responsible for which accounts, and what roles they play (Owner, Advisor, Controller, etc.), including when those responsibilities begin and end.
For this, you will want to model Parties, Accounts, and Responsibilities (with properties like role, startDate, endDate)
In LPG:
You can model this as a relationship with properties, which works until you need to connect that responsibility to something else, like a contract, policy, or data source. At that point, you’ll need to remodel it as a node with two relationships and refactor your queries.
Alternatively, you can start with a node representing the responsibility from the beginning, connected to both the party and the account. That adds an extra hop up front, but gives you more flexibility down the line.
Either way, you’re forced to make that structural decision early, before you know how the model will need to evolve.
In RDF:
You use multiple triples and supporting schema to capture role and time context. It’s expressive, but heavy to manage and slower at scale.
In DataWalk:
As shown in Figure 2, you use a Connecting Set (labeled “CS” below). This acts like a relationship for traversal but behaves like an object when needed. You can attach metadata and audit fields and link them to external references without changing the base model.
Figure 2: Connecting Set in DataWalk.
A direct comparison between an LPG and DataWalk is shown in Figure 3, below. In an LPG, you must choose between simplicity (a single direct connection at the cost of extensibility) or flexibility (adding intermediate nodes at the cost of additional hops). In DataWalk, this trade-off doesn't exist—you get simplicity and extensibility without compromise.
Figure 3: In LPG, you must choose between a connection without extendability or a set with two connections. DataWalk offers both, with one less hop as well.
Now let’s say your use case expands. You want to track financial transactions between accounts. You're modeling Transactions, with attributes like amount, timestamp, currency.
In LPG:
Once again, you must choose up front whether to model Transaction as a relationship (fast but inflexible) or as a node (flexible but more complex). If you guess wrong, you’ll need to restructure later.
In RDF:
You model each transaction using multiple triples and schema classes. You can link anything, but at a cost in volume, clarity, and performance. Figure 4 shows how RDF models can balloon when you add new components.
In DataWalk:
You create a Connecting Set: Transaction (Account → Account)
It starts simple—just a link between two accounts with properties. If you need to later link it to a fraud case or device, you can just extend it. No need to refactor the model.
Figure 4: RDF Model Expands with Addition of Transactions.
What you're modeling:
Figure 5: Fraud Investigations in DataWalk
As shown above and in the associated link chart in Figure 6, DataWalk allows you to connect one Connecting Set to another. This simplifies the analysis compared to the equivalent model in an LPG, which requires intermediate nodes.
Figure 6: In DataWalk, you can traverse from accounts to fraud investigations in just two hops. In a typical LPG model, you'd need three.
In LPG:
If you previously modeled Transaction as a relationship between accounts, you now need to remodel it as a node so it can be connected to a FraudCase. That means changing the schema, rewriting queries, and updating ingestion logic.
In RDF:
You can represent this by adding more triples and schema elements. It works—but as the model grows, so does its complexity and maintenance cost.
In DataWalk:
You simply add a new object: FraudCase, and a new Connecting Set: Investigations (Transaction → FraudCase). No existing structure needs to change. You don’t need to remodel or refactor anything. The graph just extends—cleanly and naturally.
To see how this works in practice, let’s walk through a case study where one of our customers used the model we introduced earlier—parties, accounts, responsibilities, transactions, and fraud cases. This is a financial institution that had previously invested in building out its ontology using RDF, OWL, and SHACL. While semantically rich, the previous structure had issues:
The customer adopted DataWalk and imported their existing ontology into DataWalk by ingesting RDF triples directly, including OWL and SHACL definitions. These were mapped onto DataWalk’s metamodel, preserving the original semantics without restructuring. From there:
To support the new use case, the team extended the model within DataWalk, arriving at the model shown in Figure 5 above:
With DataWalk, there were no schema rewrites and no explosion of triples. The team just naturally extended the existing model.
One of the first questions analysts needed to answer was: Which account owners are involved in a given fraud case through high-value transactions?
This analysis was built entirely in the platform using the User Interface (UI); no code or query language was required. The step-by-step traversal for this analysis is shown in Figure 8. The result was a clear, traceable view of who was involved, how they were connected, and through which accounts and transactions. More importantly, this was all done on a single pane of glass without leaving the DataWalk platform, rewriting the model, or writing custom logic.
Figure 7: First Analysis on a New Model
Here, DataWalk’s advantage becomes clear. In most RDF or LPG systems, adding a new use case like fraud detection would require early planning or significant rework. DataWalk adapts to the questions, not the other way around. The organization didn’t need to start over. They simply brought their ontology into a platform that could work with it and grow with it.
The example above shows how an existing ontology can become extensible and actionable inside DataWalk. Now let’s look at how that process works more generally.
Turning an existing ontology into a working, analytics-ready graph in DataWalk is straightforward. The process is designed to work with your existing tools and evolve with your needs.
Step | Actions | |
1 | Load Ontology Model | Import client model—e.g., OWL & SHACL definitions—directly into DataWalk. |
2 | Map to the Metamodel | Align the imported data with DataWalk’s metamodel, which defines the structure of your domain. This step transforms raw triples into a clear, navigable model tailored to your use case. |
3 | Build the Knowledge Graph | Once the metamodel is in place, DataWalk generates a fully functional knowledge graph. Entities, relationships, and connecting sets become instantly usable for search, analytics, and pattern discovery. |
4 | Refine and Iterate | Need to change your ontology? No problem—DataWalk is built for ongoing updates, whether they originate in your source system or within the platform itself. |
You can:
Whether you're adding new relationship types or reorganizing parts of the model, changes are easy to apply. You don’t need to start over. Most importantly, DataWalk and your source ontology system can update each other, enabling true two-way integration. That means users and architects stay in sync, and your knowledge graph remains a living, evolving asset.
Whether you’re reusing an existing RDF model or building something new, DataWalk gives you:
Most graph technologies provide either a model or a query engine. DataWalk provides a full platform designed to work with data end-to-end, not just store or structure it.
Key capabilities include:
These features make it easier to go from raw data to real insight, without needing separate platforms, custom scripting, or manual model rewrites.