首页> 外文会议>Offshore technology conference >Real-Time Cleaning of Time-Series Data for a Floating System Digital Twin
【24h】

Real-Time Cleaning of Time-Series Data for a Floating System Digital Twin

机译:浮动系统数字孪生的时间序列数据的实时清理

获取原文

摘要

Using accurate and high quality data is critical for any application relying heavily on the data, be it machine learning, artificial intelligence, or digital twins. Bad quality and erroneous data can result in inaccurate predictions even if the model is otherwise robust. Ensuring data quality is more critical in realtime applications where there is no human in the loop to perform sense checks on data or results. A realtime digital twin implementation for a floating system uses time-series data from numerous measurements such as wind, waves, GPS, vessel motions, mooring tensions, draft, etc. Statistics computed from the data are used in the digital twin. An extensive data checking and cleaning routine was written that performs data quality checks and corrections on the time series data before statistics are computed. Various types of errors that typically occur in a time series include noise, flat-lined data, clipped data, outliers, and discontinuities. Statistical procedures were developed to check the raw time-series for all these errors. The procedures are generic and robust so they can be used for different types of data. Some data types are slow varying (e.g., GPS) while the others are fast varying random processes. A measurement classified as an error in one type of data is not necessarily an error in the other data type. For example, GPS data can be discontinuous by nature but a discontinuity in the wave data indicates an error. Likewise, checking for white noise in mooring tension data is not that meaningful. We developed parametric data procedures so that the same routine can handle different types of data and their errors. Outlier removal routines use the standard deviation of the time-series which itself could be biased from errors. Therefore, a method to compute unbiased statistics from the raw data is developed and implemented for robust outlier removal. Extensive testing on years of measured data and on hundreds of data channels was performed to ensure that data cleaning procedures function as intended. Statistics (mean, standard deviations, maximum, and minimum) were computed from both the raw and cleaned data. Comparison showed significant differences in raw and cleaned statistics, with the latter obviously being more accurate. Data cleaning, while not sounding as high tech as other analytics algorithms, is a critical foundation of any data science application. Using cleaned time-series data and corresponding statistics ensure that a data analytics model provides actionable results. Clean data and statistics help achieve the intended purpose of the digital twin, which is to inform operators of the health/condition of the asset and flag any anomalous events.
机译:对于任何严重依赖数据的应用程序而言,无论是机器学习,人工智能还是数字双胞胎,使用准确,高质量的数据都是至关重要的。即使模型健壮,质量差和数据错误也可能导致预测不准确。在实时应用中,确保数据质量更为重要,因为在实时应用中,没有人在回路中执行对数据或结果的感觉检查。浮动系统的实时数字双胞胎实现使用来自多个测量值的时间序列数据,例如风,浪,GPS,船舶运动,系泊张力,吃水等。从数据中计算出的统计信息用于数字双胞胎中。编写了广泛的数据检查和清理例程,该例程在计算统计信息之前对时间序列数据执行数据质量检查和更正。通常在时间序列中发生的各种类型的错误包括噪声,平线数据,削波数据,离群值和不连续性。开发了统计程序来检查所有这些错误的原始时间序列。这些过程通用且健壮,因此可以用于不同类型的数据。某些数据类型变化缓慢(例如GPS),而其他数据类型变化随机过程则快速变化。在一种数据类型中被分类为错误的度量不一定在另一种数据类型中为错误。例如,GPS数据本质上可以是不连续的,但波浪数据中的不连续则表示错误。同样,检查系泊张力数据中的白噪声也没有什么意义。我们开发了参数数据过程,以便同一例程可以处理不同类型的数据及其错误。离群值删除例程使用时间序列的标准偏差,该标准偏差本身可能会因错误而产生偏差。因此,开发并实现了一种从原始数据中计算无偏统计量的方法,该方法可实现健壮的异常值去除。对多年的测量数据和数百个数据通道进行了广泛的测试,以确保数据清理程序按预期运行。统计数据(平均值,标准差,最大值和最小值)是根据原始数据和清理后的数据计算得出的。比较显示原始统计数据和清理的统计数据之间存在显着差异,后者显然更为准确。数据清理虽然听起来不像其他分析算法那么高科技,但它是任何数据科学应用程序的关键基础。使用清理后的时间序列数据和相应的统计信息可确保数据分析模型提供可操作的结果。干净的数据和统计数据有助于实现数字双胞胎的预期目的,该目的是告知操作员资产的健康状况,并标记任何异常事件。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号