We created an AI-powered knowledge layer for a multinational financial firm, unifying fragmented documentation, spreadsheets, and databases into a structured, machine-readable model. Python, Databricks, Airflow, LLMs, and RAG agents map entities, relationships, and processes across business domains. This enables fast, natural-language impact analysis, reduces dependency on key experts, and lays the foundation for future AI and analytics initiatives.

A large financial organisation operated across multiple countries, systems and regulatory environments. Over time, its knowledge landscape became fragmented:
Understanding how products, systems, data fields and processes were connected required manual document reading and expert knowledge. Impact analysis for changes (regulatory updates, system modifications, new products) was slow, risky and heavily dependent on a few key people.
The organisation needed a way to turn scattered documentation and data into a structured, AI-usable knowledge model.
We designed and implemented an AI-powered knowledge and modelling layer that automatically builds and maintains a machine-readable representation of the organisation’s business and data landscape.
The solution included:
We built data pipelines (Databricks, Apache Spark, Airflow) to ingest:
Content was cleaned, normalised and split into semantically meaningful chunks, forming the foundation of an AI-ready corpus.
Chunks were embedded and stored in vector databases, enabling retrieval-augmented generation (RAG). This allowed language models to answer questions using the institution’s own knowledge instead of generic training data.
We implemented a set of AI agents that:
The outputs were aggregated into a structured knowledge model / knowledge graph.
The institution gained:
This knowledge layer became a foundation for further AI and analytics initiatives.
< Technologies Used >



< Screenshots >