Innovative algorithms which use statistical, AI/ML, and forecasting methodologies to analyze oil-and-gas data. Note– Client data is not available on Github, and references to such data has beeen de-identified for security purposes
- Dependencies
- Notable File Structure
- Design Decisions
- Pipeline
- Benchmarking and Backtesting
- Final Implementation
- Other
Dependency Chart for entire repository
File structure content here
Important design decisions and their motivations
There are seven sequential components in the main pipeline of this repository.
- Source Table Aggregation
- Physics Feature Engineering
- Anomaly Detection and Aggregation
- Mathematical Feature Engineering
- Model Creation and Performance
- Optimization and Producer Level Steam Allocation
- Injector Level Steam Allocation
Process Flow of Pipeline
Joining the production, production testing, injection, and fiber data.
Generating new features based on domain-specific [modified] SAGD knowledge.
Using custom anomaly detection package to determine weights.
Generating new features - solely based on combinatorial explosions that are helpful towards model performance.
Running the h2o experiment, assesssing model performance, and picking the best ones.
Determinining optimization base table and using domain-specific contraints to tune high-level steam allocation.
The highest resolution of steam allocation.
Using the top performing models and running incremental backtests to compare recommended and actual steam allocations.
Based on model performance and general design of project, what's the best way to implement it.
Miscellaneous.