Madhuri K
Senior Data Modeler
Email:
[email protected] | Phone: +1 (972) 853-1869
PROFESSIONAL SUMMARY:
Senior Data Modeler with over 10 years of progressive IT experience in data modeling, data
management, ETL development, data warehousing, and analytics across OLTP and OLAP systems in
healthcare, finance, insurance, and enterprise domains
Expert in multidimensional modeling using Kimball's approach, specializing in star schema and
snowflake schema design, supporting both traditional and cloud-based architectures (Google
BigQuery, Snowflake, Oracle).
Designed and implemented atomic-level transaction fact tables for high-precision reporting and SCD
Type 2 tracking to ensure accurate historical analysis.
Proficient in 3rd Normal Form (3NF) modeling for OLTP systems, ensuring data consistency,
normalization, and referential integrity across large-scale ERP applications including UKG and Oracle
HCM.
Extensive hands-on experience with ER/Studio and Erwin Data Modeler, performing conceptual,
logical, and physical data modeling, model versioning, reverse engineering, and forward engineering
across multiple platforms.
Designed canonical data models, metadata repositories, and reference data frameworks for
enterprise-scale systems, facilitating seamless integration, governance, and semantic alignment.
Led enterprise data management initiatives to define standard naming conventions, enforce data
quality, and ensure model consistency across data products and reporting systems.
Built and maintained master data management (MDM) frameworks for employee, department, payroll,
and location domains, aligning with enterprise data governance policies and ensuring single source of
truth.
Deep experience modeling structured, semi-structured, and unstructured data in cloud data lakehouse
environments, integrating tools like Hive, Hue, S3, and Databricks.
Supported metadata-driven development by documenting data lineage, business glossary definitions,
source-to-target mappings, and transformation rules to improve traceability and auditability.
Created high-performance BigQuery schemas using clustering, partitioning, and materialized views to
optimize query performance and reduce cloud spend.
Designed and modeled enterprise data warehouses and analytical solutions using Kimball dimensional
modeling principles for multiple business domains including HRIS, Payroll, Timekeeping, and
Compliance.
Proficient in designing both OLTP and OLAP data structures, translating complex business
requirements into scalable and high-performing data architectures.
Collaborated with cross-functional teams to design data dictionaries, entity-relationship diagrams
(ERDs), and functional specs that enable smooth coordination across analytics, BI, and engineering
teams.
Integrated SAS, Power BI, and Tableau with modeled datasets to support advanced analytics and
regulatory reporting needs.
Skilled in ETL tools such as Informatica Cloud, SSIS, and DataStage, building and optimizing pipelines
for high-volume, real-time, and batch processing.
Experience in RDBMS development using Oracle PL/SQL, SQL Server, and Teradata, including stored
procedures, performance tuning, indexing, and query optimization for large-scale systems.
Strong grasp of data governance, data quality, data profiling, and data privacy principles including
compliance with HIPAA, GDPR, and internal standards.
Adept at migrating legacy systems to modern cloud platforms (GCP, AWS), performing data mapping,
transformation logic reengineering, and schema rationalization for scalable architecture.
Developed insurance-specific data models aligned with industry frameworks (e.g., ACORD, IBM
Insurance Data Warehouse), ensuring semantic consistency and standardization across data sources.
Modeled complex NoSQL and semi-structured data structures (e.g., JSON, Avro, Parquet) using
Hackolade, ensuring alignment with business and domain-driven design standards.
Applied data profiling techniques to identify data quality issues and implemented cleansing rules to
improve downstream analytics accuracy and reliability.
Built entity-relationship diagrams (ERDs) and data dictionaries in Hackolade to facilitate
communication with data architects, engineers, and business stakeholders.
Created data flow documentation, entity relationship diagrams (ERDs), and functional specifications to
support cross-team development and data literacy.
Used Teradata Fast Export, BigQuery export tools, and federated queries for high-volume data
extraction and business reporting workloads.
Designed and implemented robust data models for financial systems, ensuring efficient storage,
retrieval, and analysis of patient data while adhering to HIPAA regulations.
Experienced in maintaining data quality, collecting requirements, data analysis, architecture design,
and developing data models, ETL processes, and functional specifications.
Migrated legacy database applications to newer technologies, performing data movement, mapping,
and cleansing.
TECHNICAL SKILLSET:
Database Specialties Database Architecture, Data Analysis, Enterprise Data Warehouse,
Database Design and Modeling, Multidimensional Modeling, Star/Snowflake
Schema Design, Data Vault 2.0, Metadata Management, Reference Data
Modeling, Master Data Management (MDM), Data Governance, Data
Integration and Migration, ETL Architecture and Design, OLTP, OLAP, Data
Lakehouse Architecture, Cloud Data Warehousing.
Databases MySQL, SQL Server 2019/2016/2012/2008, Oracle 19c/12c/11g/10g,
Teradata, DB2, Snowflake, Google BigQuery, MongoDB, Netezza.
Programming & Query SQL, PL/SQL, T-SQL, Python (for data processing and scripting), UNIX Shell
Languages: Scripting, PERL.
Operating Systems Windows Server (2016/2019), UNIX/Linux, MS DOS, Sun Solaris.
Reporting & BI Tools: Tableau, Power BI, Crystal Reports XI, SSRS, SAP BusinessObjects (5.x /
6.x / XI), Looker
ETL & Data Integration Informatica PowerCenter, Informatica Cloud (IICS), IBM DataStage, Ab Initio,
Tools: Talend, SSIS, Data Junction, Oracle Data Integrator (ODI), Airflow
Relational Databases Oracle, MySQL, PostgreSQL, Microsoft SQL Server, SQLite NoSQL Databases
MongoDB, Cassandra, Couchbase, Redis, Amazon DynamoDB.
Web Technologies HTML5, DHTML, XML, XSD, JSON, CSS.
Scripting Languages VBScript, JavaScript, Python, UNIX Shell Scripting
Cloud & Big Data AWS (S3, RDS, Redshift, Glue), Google Cloud Platform (BigQuery, Dataflow,
Platforms Pub/Sub), Azure Data Factory, Databricks, Hadoop Ecosystem (Hive, Hue),
Snowflake
Project Execution & Data Ralph Kimball and Bill Inmon Methodologies, Data Vault 2.0, IBM UDMH,
Methodologies Rapid Application Development (RAD), Joint Application Development (JAD),
Rational Unified Process (RUP), Agile/Scrum, SDLC.
Other Tools & Utilities TOAD, BTEQ, TPT, Fast Load, MultiLoad, FastExport, SQL Developer, Git,
JIRA, Confluence, MS Office Suite (Word, Excel, PowerPoint, Outlook, MS
Project).
Modeling Tools Erwin 9.x, ER/Studio, PowerDesigner, Embarcadero, Hackolade, MS Visio,
Rational Rose
WORK EXPERIENCE:
Charles Schwab - Dallas, Tx Jul 2024 - PRESENT
Role: Senior Data Modeler
Responsibilities:
Built dimensional models with multiple fact tables (transactional, periodic snapshot, accumulating
snapshot) and conformed dimensions to support integrated analytics across business lines.
Collaborated with BI and reporting teams to ensure model designs supported OLAP cube
development, drill-down analysis, and cross-subject area reporting.
Designed and enforced entity naming standards, data types, domains, and abbreviations in ER/Studio
and Erwin, ensuring consistency across models and layers.
Performed impact analysis for schema changes across upstream systems, ETL workflows, reports, and
APIs to maintain model integrity during release cycles.
Defined primary, foreign, and surrogate key strategies, cardinality, optionality, and indexing for
performance tuning in Oracle, Redshift, and BigQuery environments.
Created ER diagrams, relationship matrices, and model documentation for stakeholders using
ER/Studio and published versions via model repositories (Model Mart).
Modeled hierarchical reference data (e.g., organization units, product categories, geo structures) and
enforced controlled vocabularies and codes via lookup entities.
Integrated metadata layers into Power BI and Tableau for semantic alignment, enabling business users
to discover and consume enterprise data consistently.
Worked closely with data stewards and compliance teams to design GDPR/PHI-compliant models,
implementing data masking, data minimization, and lineage tracking.
Architected subject-oriented data marts aligned with functional domains (e.g., sales, finance,
operations) to serve performance-optimized analytical needs.
Implemented bi-temporal modeling for insurance and healthcare systems to handle as-was and as-is
data for compliance and legal audit reporting.
Used ER/Studio’s repository collaboration features (Team Server, naming standards, versioning) to
manage distributed data modeling efforts across global teams.
Validated data completeness and transformation rules through sample profiling and row-level lineage
tracing as part of modeling QA.
Coordinated with DevOps and DBA teams to automate DDL deployment and synchronization scripts
across environments using CI/CD tools like Azure DevOps and Bitbucket pipelines.
Evaluated and selected modeling patterns (e.g., hybrid star-snowflake, hub-spoke MDM, data vault)
based on business use case, performance profile, and regulatory requirements.
Modeled time-series, hierarchical, and semi-structured data (JSON/XML) to support modern data lake
and NoSQL ingestion scenarios using Hackolade and native DDL extensions.
Supported downstream ML/AI initiatives by designing feature-store-ready analytical base tables
(ABTs) with clear dimensional keys and granularity alignment.
Maintained CI/CD pipelines for ETL workflows using Informatica Cloud, managed deployments, and
automated monitoring and alerting mechanisms for production stability.
Developed a high-level conceptual model using Erwin based on gathered requirements.
Integrated XML and XSD into ETL workflows using Informatica and Talend for data transformation.
Contributed to data mapping activities for efficient data integration within the data warehouse.
Provided detailed source-to-target mappings for the ETL team to facilitate data loading into the data
mart (initial, full, and incremental).
Led the migration of data and data models from an SQL Server environment to Oracle 19c.
Environment:
ER/Studio, Oracle SQL Developer Data Modeler, Hackolade, Microsoft SQL Server 2022 (SSAS, SSIS, SSRS),
Oracle Database 19c, AWS Redshift, Google BigQuery, DB2 LUW, Teradata Vantage, Informatica PowerCenter
10.x, Informatica Intelligent Cloud Services (IICS), Talend Open Studio, IBM DataStage 12.0, SAP
BusinessObjects BI 4.3, SAS Enterprise Guide, SAP HANA (Modeling Views), Power BI, Tableau, IBM Rational
Rose 9.7, JIRA, Confluence, Git, Bitbucket, Microsoft Azure DevOps, TOAD for Oracle, SQL Developer, Model
Mart, Microsoft Office 365.
BlueCross BlueShield – DALLAS, TX Aug 2022 - Jun 2024
Senior Data Modeler
Responsibilities:
Designed and maintained logical and physical dimensional data models using Erwin Data Modeler,
generating DDL scripts and collaborating with DBAs to create tables, views, primary/foreign keys, and
indexes in Oracle and SQL Server databases.
Built star and snowflake schemas for reporting and data marts in P&C insurance, enabling KPI tracking
and compliance analytics across underwriting, claims, and billing domains.
Migrated legacy SSIS-based ETL pipelines to modern ELT architecture in Snowflake, optimizing
transformation performance through pushdown SQL logic.
Translated complex data requirements into scalable models by working with actuarial, underwriting,
finance, and claims teams, aligning designs to business domain logic.
Applied 3rd Normal Form (3NF) modeling principles for OLTP and denormalized OLAP structures for
high-performance analytical queries.
Developed canonical models to harmonize source system data structures and enhance reusability
across multiple ETL and reporting layers.
Analyzed OLTP source systems and ODS layers to identify relevant entities and attributes, and
constructed metadata mappings to facilitate accurate data integration.
Created domain-specific data models following DDD (Domain-Driven Design) concepts to support
modular design and data mesh strategies.
Designed data models adhering to General Electric’s Crow’s Foot notation, including atomic transaction
fact tables and conformed dimensions for cross-functional analytics.
Led the design of data marts with over 10+ tables, utilizing both star and snowflake schema
methodologies tailored to business subject areas.
Provided data modeling governance by defining naming conventions, attribute definitions, and
documentation standards in Model Mart and Erwin repositories.
Assisted developers and DBAs with model interpretation and translation, bridging business definitions
with technical implementation.
Designed PL/SQL stored procedures, triggers, and database security layers, managing user access and
data loading through Oracle utilities.
Developed ETL mappings in Informatica to load data from Oracle, SQL Server, DB2, flat files, and Excel,
employing transformations and filters for data cleansing and conformance.
Enabled metadata-driven data integration by aligning source-to-target mappings, lineage
documentation, and entity traceability across systems.
Built and integrated reference data entities and lookups across systems to ensure consistency in
business keys and reduce data quality issues.
Maintained detailed SDLC documentation, including system specs, data dictionaries, model annotations,
and test plans for regulatory audit readiness.
Created automated workflows to extract data from Oracle, SQL Server, flat files, and Excel, supporting
ingestion into enterprise data lakes and marts.
Collaborated with business stakeholders to define analytical reporting requirements and translated
them into performant and scalable physical models.
Environment:
Erwin Data Modeler v22.0, Oracle 19c, Hackolade, SQL Server 2022 (SSIS/SSRS), DB2 LUW, Informatica
PowerCenter, Informatica Intelligent Cloud Services (IICS), SAS, Oracle Developer Suite, Toad for Oracle,
PL/SQL Developer, Model Mart, Apache, Excel/Flat Files, Windows Server, UNIX/Linux.
Voyager - Dallas, TX Jul 2019 to Jul 2022
Data Modeler/ Data Analyst
Responsibilities:
Engaged with business users to analyze end-to-end processes and transform functional requirements
into conceptual, logical, and physical data models, driving robust database design and enterprise-wide
standardization.
Participated in JAD (Joint Application Design) sessions to gather business needs and align data
modeling initiatives with enterprise financial and operational objectives.
Created and optimized logical and physical data models using Erwin Data Modeler, enforcing
referential integrity and normalization across subject areas.
Defined canonical entity models and standardized mappings to unify data definitions across domains,
enabling consistency across APIs, ETL, and reporting layers.
Applied both Kimball and Inmon methodologies to architect flexible and scalable data warehouse
solutions, selecting modeling strategies based on project-specific analytics, compliance, or
performance needs.
Supported legacy-to-cloud modernization by integrating data from legacy systems into modern
platforms like Snowflake, Databricks, and Azure Synapse, preserving business logic and lineage.
Designed star and snowflake schemas to support business intelligence and analytics within the P&C
insurance domain, including identification of facts, dimensions, and hierarchies.
Created physical data models (PDMs) for OLAP systems using ER/Studio and collaborated with DBAs
to implement schema designs and performance-optimized DDLs.
Led initiatives on Data Governance, designing data transformation, standardization, and cleansing
rules to support MDM architecture across OLTP, ODS, and OLAP systems.
Built and validated source-to-target mappings for ETL pipelines in collaboration with Informatica and
BI teams, ensuring data integrity across ingestion and transformation layers.
Modeled real-time event-driven data flows by designing Kafka topics and Avro schema visualizations
using Hackolade, enabling seamless stream processing integration.
Coordinated with ETL, BI, and DBA teams to identify and resolve data quality issues, improve
transformation logic, and validate schema compatibility across data warehouse layers.
Authored and maintained a comprehensive data dictionary and metadata repository, supporting both
technical documentation and business stakeholder reporting needs.
Designed data models tailored for legal industry compliance, focusing on high data accuracy, auditable
structures, and secure entity definitions.
Understood and applied business analysis principles to model data flows, define transformation logic,
and optimize database structures for transactional and analytical workloads.
Utilized Hackolade to visualize Kafka event structures and Avro schemas, enabling real-time data
stream modeling in event-driven architectures.
Worked with ETL to create source-to-target mappings and performed validation for source-to-target
mappings.
Collaborated with ETL, Bl, and DBA teams to analyze and provide solutions to data issues and other
challenges while implementing the OLAP model.
Developed and maintained a data dictionary to create metadata reports for technical and business
purposes.
Developed and maintained data models tailored to the legal industry, ensuring compliance and
accuracy in reporting practices.
Understood basic business analysts' concepts for logical data modeling, data flow processing, and
database design.
Environment:
Erwin 2016, Toad, PL/SQL, Oracle 12c, Hackolade, SQL Server 2016, Windows Server 2016, Quest Central for
DB2 v 2016, COBOL, Teradata, Microsoft SQL Server 2016 Reporting Services
AT & T, Cerritos, California Nov 2017 – Jun 2019
Role: Data Modeler/ Data Analyst
Responsibilities:
Worked as OLTP Data Architect & Data Modeler to develop the Logical and Physical 'Entity
Relational Data Model' for Claim system (Claim & Adjustments) with entities & attributes and
normalized them up to 3rd Normal Form using ER/Studio.
Involved in creating Physical and Logical models using Erwin.
Collaborated with business analysts and architects to ensure canonical models aligned with business
semantics and operational needs.
Created and maintained Database Objects Tables, Views, Indexes, Partitions, Synonyms, Database
triggers, and Stored Procedures in the data model.
Designed the ER diagrams, logical model relationship, cardinality, attributes, candidate keys and
physical database capacity planning, object creation, and aggregation strategies for Oracle and
Teradata as per business requirements using Erwin.
Extracted data from various sources like Oracle, Mainframes, and flat files and loaded them into the
target Netezza database.
Ensured models aligned with compliance frameworks (e.g., SOX, HIPAA for claims), and contributed to
internal data risk assessments.
Extensively used ETL methodology for supporting data extraction, transformations, and loading
processing, in a complex EDW using Informatica.
Responsible for analyzing various data sources such as flat files, ASCII Data, EBCDIC Data, Relational
Data Oracle, DB2 UDB, and MS SQL Server from various heterogeneous data sources.
Experience in Data Transformation and Data Mapping from source to target database schemas and
data cleansing.
Involved in extensive DATA validation using SQL queries and back-end testing.
Used SQL for Querying the database in a UNIX environment.
Written complex SQL queries for validating the data against different kinds of reports generated by
Business Objects XIR2.
Create tables and queries to produce additional ad-hoc reports.
Performed data mining on Claims data using very complex SQL queries and discovered claims patterns.
Used Teradata OLAP used like RANK, ROW NUMBER, QUALIFY, CSUM, and SAMPLE.
Designed and developed cubes using SQL Server Analysis Services SSAS using Microsoft Visual Studio.
Environment:
PL/SQL, Business Objects 4.2, ETL Tools Informatica 10, Oracle 11g, Teradata 15, Teradata SQL Assistant 15,
Netezza, Power Designer 16.5, Erwin 2016, Informatica MDM 10.
Client: Despot Power Systems, Hyderabad, India Jun 2015 – Oct 2017
Role: Data Modeler
Responsibilities:
Collaborated in designing databases with a focus on tightly bound Key constraints, ensuring optimal
data relationships and integrity.
Worked on building up the data model for Mobile Banking in the data warehouse.
Worked on designing a Relational Model and Dimensional Model for Interactive Services such as
Internet Banking, PC banking, and Bill Pay for the bank.
Worked on designing the Net Loss Model for the bank which involves data regarding loans and
recoveries.
Worked on building a conceptual and logical model starting from a basic field analysis spreadsheet of
Credit Mart as a part of one of my projects.
Used Graphical Entity-Relationship Diagramming to create a new database design via an easy-to-use,
graphical interface.
Developed logical data models and physical data models that confine existing condition/potential
status data fundamentals and data flows using ER-Studio.
Defined the Primary Keys PKs and Foreign Keys FKs for the Entities, created dimensions model star and
snowflake schemas using Kimball methodology.
Closely worked with business analysts/Data Modeler in the process of requirements gathering for
various application-specific databases.
Facilitated JAR/JAD sessions to determine data definitions and business rules governing the data and
facilitated review sessions with subject matter experts SMEs on the logical and physical data models.
Designing and customizing data models for Data warehouse supporting data from multiple sources in
real-time.
Worked on Metadata Repository MRM for maintaining the definitions and mapping rules up to mark.
Applied data naming standards, created the data dictionary documented data model translation
decisions, and maintained DW metadata.
Used data profiling tools and techniques to ensure data quality for data requirements.
Environment:
ER/Studio 2015, IBM Infosphere DataStage 11.5, Microsoft SQL Server 2014, IBM Data Studio 4.1, EMC
Greenplum 4.3, Oracle 12c, IBM DB2 10.5, Business Objects 4.1, pgAdmin 4