Common questions and answers about using the IIoT Data Quality Assessment Service.
The IIoT Data Quality Assessment Service is a web application for analyzing and assessing the quality of high-frequency sensor data from Industrial Internet of Things (IIoT) devices. It provides tools for data import, visualization, quality assessment, and AI-powered insights.
- Data Import: Upload and import CSV sensor data files
- Data Visualization: Interactive charts and advanced analytics
- Missing Values Analysis: Identify and analyze missing data
- Invalid Values Analysis: Detect threshold violations
- Data Quality Assessment: Comprehensive quality metrics
- DQA Agent: AI-powered chat assistant
The application works best with:
- Chrome/Edge (latest versions)
- Firefox (latest versions)
- Safari (latest versions)
For end users, you only need:
- Docker and Docker Compose (for services)
- Node.js 18+ (for frontend)
- A web browser
The backend runs in Docker containers, so Python is not required for end users.
See the Getting Started Guide for detailed installation instructions. The quickest way is:
- Clone the repository
- Create
.envfile fromenv.example - Add your OpenAI API key (for DQA Agent)
- Run
./start-dev.sh
The OpenAI API key is required for the DQA Agent feature, which provides AI-powered chat assistance. Other features work without it.
Run the startup script:
./start-dev.shThis starts all services automatically.
Run the stop script:
./stop-dev.shOr manually:
docker-compose downThe application supports CSV files for:
- Sensor data (with timestamp column)
- Sensor metadata/tags (with required columns)
See Data Import Guide for detailed format requirements.
The data file must have:
- A
timestampcolumn (first column) - One column per sensor tag
The tags file must have these columns:
- TAG
- Tag Description
- MACHINE_GROUP
- LOW_THRESHOLD
- HIGH_THRESHOLD
- THRESHOLD_TYPE
- AGGREGATION_RULE
- ENGINEERING_UNITS
- CATEGORY
Yes, you can specify a comma-separated list of sensor tags during import. Leave empty to import all sensors.
Import time depends on data size:
- Small datasets (< 10,000 rows): 1-2 minutes
- Medium datasets (10,000 - 100,000 rows): 5-10 minutes
- Large datasets (> 100,000 rows): 10+ minutes
Check the Troubleshooting Guide for import issues. Common causes:
- File format errors
- Database connection issues
- Insufficient disk space
- Go to Data Loading page
- Select a machine group from the dropdown
- Data loads automatically
See Data Loading Guide for details.
- Raw Data: Original sensor readings as imported
- Preprocessed Data: Aggregated data based on aggregation rules (min, max, avg, sum)
Many pages support date range filtering:
- Set Date From and Date To
- Click Apply Filter
- Analysis updates for selected range
Export options vary by page:
- Screenshots: Use browser tools
- Data export: Available on some pages (CSV format)
- Reports: Generate quality reports
Data quality refers to the fitness of data for its intended use, measured by:
- Completeness: How much data is present
- Accuracy: How correct the data is
- Consistency: How uniform the data is
The system evaluates:
- Missing value percentages
- Threshold violations (accuracy)
- Duplicate records (consistency)
- Outlier detection
- Correlation analysis
See Data Quality Guide for details.
Quality levels:
- Excellent: All metrics within acceptable ranges
- Good: Minor issues, generally acceptable
- Fair: Some concerns, monitor closely
- Poor: Significant issues, take action
- Identify specific issues (missing values, accuracy, etc.)
- Investigate root causes
- Address sensor hardware issues
- Check data collection system
- Re-assess after improvements
Missing values are expected sensor readings that are not recorded. They can occur due to:
- Sensor malfunctions
- Data collection interruptions
- Transmission failures
The system compares expected readings (based on timestamp frequency) with actual readings to identify gaps.
Guidelines:
- < 5%: Excellent
- 5-10%: Good
- 10-20%: Moderate concerns
- > 20%: Poor, take action
- Investigate sensor hardware
- Check data collection system
- Review transmission logs
- Consider sensor replacement if needed
Invalid values are sensor readings that violate defined thresholds:
- Below low threshold (for "Down" type)
- Above high threshold (for "Up" type)
- Outside range (for "Up/Down" type)
Thresholds are defined in the tags metadata file:
- LOW_THRESHOLD: Lower limit
- HIGH_THRESHOLD: Upper limit
- THRESHOLD_TYPE: Up, Down, or Up/Down
Guidelines:
- < 5%: Good data quality
- 5-15%: Some concerns, investigate
- > 15%: Significant issues, take action
- Check sensor calibration
- Verify threshold settings
- Review sensor hardware
- Check operating conditions
The DQA Agent is an AI-powered chat assistant that helps you:
- Ask questions about data quality
- Get insights and recommendations
- Understand quality metrics
- Troubleshoot issues
Yes, the DQA Agent requires a valid OpenAI API key. Other features work without it.
You can ask about:
- Data quality metrics
- Specific sensors
- Recommendations
- Metric explanations
- Troubleshooting help
See DQA Agent Guide for examples.
Agent responses are AI-generated and should be verified. Use as guidance, not absolute truth. Cross-check important findings with other analyses.
- Check Docker is running:
docker info - Verify ports are not in use
- Check
.envfile exists and is configured - Review startup logs:
docker-compose logs
- Check if frontend is running:
lsof -ti tcp:5173 - Verify Node.js is installed:
node --version - Check for port conflicts
- Review browser console for errors
- Check backend logs:
docker-compose logs backend - Verify database connection
- Check backend health:
curl http://localhost:8000/health - Restart backend:
docker-compose restart backend
- Check TimescaleDB is running:
docker-compose ps timescaledb - Verify environment variables
- Check database logs:
docker-compose logs timescaledb - Test connection:
docker-compose exec timescaledb pg_isready
See Troubleshooting Guide for more solutions.
- Reduce date range in filters
- Select fewer sensors
- Use aggregated data source
- Check system resources
- Close unnecessary browser tabs
- Use date range filters
- Select fewer sensors
- Wait for charts to fully load
- Check data volume
Large files take time. Monitor progress and be patient. Consider importing subset of sensors first.
- Always validate files before importing
- Start with small test dataset
- Backup existing data before overwriting
- Verify file encoding (UTF-8)
- Check sensor tags match between files
- Start with overview (all sensors)
- Focus on specific sensors if needed
- Use date filters for large datasets
- Compare results across time periods
- Document findings
- Assess quality regularly
- Set quality standards
- Monitor trends over time
- Take action on identified issues
- Track improvements
- Documentation Index - Overview of all guides
- Getting Started Guide - Setup instructions
- Troubleshooting Guide - Common issues
- Main README - Project overview
Check the project repository for issue reporting guidelines. Include:
- Description of the issue
- Steps to reproduce
- Error messages
- System information
Check the project repository for contribution guidelines and code of conduct.
- Getting Started Guide - Installation and setup
- Data Import Guide - Importing data
- Data Quality Guide - Quality assessment
- Troubleshooting Guide - Problem solving
Have a question not answered here? Check the Troubleshooting Guide or consult the specific feature guides.