You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Installer for DataKitchen's Open Source Data Observability Products. Data breaks. Servers break. Your toolchain breaks. Ensure your team is the first to know and the first to solve with visibility across and down your data estate. Save time with simple, fast data quality test generation and execution. Trust your data, tools, and systems end to end.
DataOps Data Quality TestGen is part of DataKitchen's Open Source Data Observability. DataOps TestGen delivers simple, fast data quality test generation and execution by data profiling, new dataset hygiene review, AI generation of data quality validation tests, ongoing testing of data refreshes, & continuous anomaly monitoring
CSV Data Validator is a tool to validate csv file. It parse csv and validate the data with .hdr(csv meta data) before ingestion to Data Lake. It checks data file availability for every day load and validate data with respective meta data like File Size, Checksum, Delimiter, Record count etc. It ensure landed data conformity before give go ahead …
This repository showcases my work in a Data Analytics and Commercial Insights simulation. Tasks include data preparation, customer analytics, and uplift testing using transaction data to generate strategic, data-driven recommendations. Outputs include code, benchmark analysis, and reports aimed at supporting informed business decisions.
The main purpose of this repository is to build the pipeline for training of regression models and predict the compressive strength of concrete to reduce the risk and cost involved in discarding the concrete structures when the concrete cube test fails.
Created a high-performance REST API using FastAPI and Pydantic, supporting data on various countries, with features such as data validation and automatic JSON response handling.
🎯 Gender inequality at work - use of KNIME (Background research, GDPR, Data governance, ETL, EDA, Data cleaning and validation, Statistical tests with R, and Data Visualization)