Business AnalysisBusiness Analyst

What methodology would you employ to decommission a mission-critical shadow IT **Microsoft Access** database that underpins a monthly financial close process worth $2M in working capital decisions, when the business unit vehemently opposes migration to the corporate **SAP** **ERP** citing a 300% speed advantage, the database contains unmasked **PII** subject to **GDPR** Article 33 breach notification requirements, and the original developer departed six months ago leaving zero documentation?

Pass interviews with Hintsage AI assistant

Answer to the question

The methodology centers on risk-mitigated coexistence followed by strangulation. Rather than big-bang migration, you establish a data governance layer that extracts, transforms, and validates Access data in real-time while shadowing the ERP workflow. This creates an auditable bridge that satisfies compliance immediately while proving functional parity through A/B testing.

Situation from life

At a mid-sized retail conglomerate, the Treasury department relied on a decade-old Access application to calculate cash-flow forecasts every month-end. The application queried seventeen disparate Excel files and a legacy AS/400 terminal screen scrape, completing the close in four hours versus the SAP module's twelve-hour runtime. When the GDPR compliance audit flagged unencrypted customer payment terms stored in local tables, the CFO mandated remediation within 90 days, but the Treasury VP threatened to resign if workflow speed degraded.

Three solutions emerged for board consideration. The first proposed immediate hard-cut migration to SAP, arguing that regulatory risk outweighed user convenience. This offered immediate compliance and single-source truth, but carried catastrophic operational risk: the SAP module lacked support for the proprietary allocation algorithms embedded in the Access VBA macros, guaranteeing a failed month-end close and potential liquidity crisis that could freeze vendor payments.

The second suggested rebuilding the logic in a modern Python/Django web application with a PostgreSQL backend. This promised perfect feature replication and cloud scalability, but required six months of development—exceeding the compliance deadline—and introduced new infrastructure costs without addressing the immediate GDPR exposure or user training requirements.

The third solution, selected after intensive stakeholder workshops, implemented a Microsoft Power Automate extraction layer that sanitized PII through deterministic tokenization before writing to a compliant Azure SQL data warehouse. The Access frontend remained temporarily intact for user interaction, but all data persistence redirected to the encrypted warehouse, creating a hybrid where Treasury retained their processing speed while GDPR requirements were technically satisfied. A parallel track began translating VBA logic into SAP ABAP routines using recorded user sessions as pseudocode references.

The result achieved compliance on day 87 without disrupting the close process. Six months later, the SAP module achieved functional parity through iterative refinement guided by the tokenized dataset, allowing the Access database to retire gracefully with zero business downtime.

What candidates often miss

How do you calculate the precise technical debt cost of maintaining the shadow system versus migration when the business refuses to quantify "speed" in monetary terms?

Candidates often fail to translate qualitative user experience into financial risk metrics. You must model the shadow IT cost as the sum of potential compliance violation penalties (4% global turnover under GDPR), the actuarial cost of the single point of failure (probability of database corruption × cost of missed financial filings), and opportunity cost of IT support hours diverted to maintaining legacy technology. Present this as a monthly "risk rental fee" that the business is effectively paying to avoid change, making the abstract concrete for executives.

What specific data lineage techniques would you apply when the Access database contains calculated fields with no formula documentation and circular references between tables?

Most candidates suggest manual inspection or user interviews, which is insufficient for a complex Access application. The correct approach involves automated schema discovery tools like Microsoft Access Analyzer or ApexSQL to reverse-engineer table relationships, coupled with runtime tracing using ODBC query logging to capture actual execution paths during the month-end process. For calculated fields, you export all VBA modules into text and parse for assignment patterns using regex, then cross-reference with front-end form controls to distinguish display formatting from actual business logic. This creates a deterministic map of data lineage without relying on tribal knowledge.

How do you structure the governance transition to prevent the business unit from simply recreating the same shadow IT problem six months later in Power BI or similar self-service tools?

Candidates miss the sociotechnical dimension of shadow IT proliferation. The solution requires establishing a "citizen developer" governance charter that permits agile self-service within technical guardrails. Implement a Data Loss Prevention (DLP) policy on all corporate endpoints that blocks local storage of sensitive data categories, forcing usage of approved cloud repositories with audit trails. Simultaneously, create a fast-track DevOps pipeline where business units can request approved datasets with a 48-hour SLA, eliminating the latency that originally drove them to shadow IT. Without fixing the demand-side frustration through service improvement, technical controls merely displace the problem to another ungoverned tool.