• 936-336-5652

    Archives for October 2025

    Sendero Midstream

    SCADA Architecture for Multi-Site Saltwater Disposal (SWD) Operations

    Produced-water volumes keep climbing, regulations keep tightening, and running each SWD site as a standalone island creates blind spots you can feel in uptime, trucking logistics, and compliance. 

    A well-designed SCADA (Supervisory Control and Data Acquisition) system becomes the nerve center: one view, many sites, consistent decisions. It centralizes monitoring (injection rates, pressures, equipment health) and compliance evidence while allowing local control to ride through communication hiccups. 

    Security and reliability patterns for these systems are well-documented in NIST SP 800-82 (Rev. 3) and the ISA/IEC 62443 family. Use them as the backbone for design choices.

    Cross-Site Scheduling for Smarter Operations

    Centralization pays off when wells hit capacity unevenly, county rules don’t match, and trucks queue at the wrong gate. With site-to-site visibility and cross-site scheduling, you can smooth loads, redirect trucks, and tighten reporting. This is especially useful with the strict injection limits and integrity testing that regulators emphasize.

    The EPA’s UIC Class II program frames the big picture. Day-to-day limits and forms are set by state agencies, like Texas RRC, Oklahoma OCC, New Mexico OCD.

    Field Realities You Have to Design Around

    Distance and Networks

    Remote SWD sites rarely get high-speed fiber optic network connectivity. You’ll juggle cellular, licensed/unlicensed radio, and satellite, each with their own latency and availability trade-offs.

    Architect for graceful degradation: keep essential control local and push summaries to the center. 

    Regulations Vary

    Texas commonly conditions injection permits with a maximum surface injection pressure of about 0.5 psi per foot to the top of the injection interval. A practical ceiling intended to avoid fracturing confining zones. Oklahoma and New Mexico impose different pressure/testing particulars and reporting cadences. Every new field brings different regulations and that can make centralized accounting and regulatory reporting seem impossible

    A centralized SCADA environment can easily bridge all the requirements of the complex realities of a large corporate structure. Whether it’s templating reports per state so operators aren’t word-smithing spreadsheets at midnight or post-processing field data to aggregate accounting needs, a SCADA system gives businesses the agility to compete.

    Capacity Balancing

    Without system-wide visibility, one site hits its daily limit while another idles. Central dispatch guided by historian trends and real-time KPIs (injection efficiency, uptime, alarms) curbs wasted trucking miles and improves compliance headroom. 

    Manipulate water traffic based on available capacity and operational needs. Centralizing the data allows decision making based on real-time data of a system, as a whole, not in a vacuum.

    Safety and Environmental Signals

    You’re watching formation pressures, permitted rates, water quality, and leak/spill indicators continuously. Hitting limits isn’t optional, it’s the line between steady operations and citations.

    What to Monitor And Why

    Pressures and rates. They define safe operating envelopes and permit compliance. Deviations trigger operator notifications.

    Water quality. Salinity, oil/solids carryover, and treatment efficacy influence disposal formation compatibility and maintenance cycles.

    Equipment health. Use vibration/temperature/runtime to drive condition-based maintenance so a failing pump doesn’t become a shutdown.

    Data harmonization. Different pads run on mixed protocols (EtherNet/IP, Modbus, legacy RTUs), so standardizing tags/units is critical. DNP3 suits unreliable links with event reporting, while OPC UA offers secure, interoperable data modeling for modern systems.

    Cybersecurity isn’t optional. Treat all SCADA systems as critical infrastructure: zone/segment by function, route traffic through conduits, apply least privilege, and instrument for detection.

    Core Architecture for Multi-Site SWD

    Central Servers, Thoughtfully Redundant

    Keep primary and standby in separate locations; use clustering for historians/alarms so a single failure doesn’t blank your visibility. This mirrors OT guidance for resilience rather than fragile, single-box “do-everything” servers.

    Operator Interfaces that Scale

    Start with a map-level overview for status at a glance, click into facility screens for that site’s specific equipment and control, and standardize navigation so an operator can cover all of their facilities without relearning screen logic at each site.

    Rugged Field Controllers

    PLCs/RTUs must survive heat, cold, dust, and vibration. Outdoor enclosures rated NEMA 4X protect against hose-down, wind-blown dust, and corrosion.

    Hazardous areas typically call for Class I, Division 2-appropriate equipment selection and installation.

    Protocol Mix

    A SCADA environment allows you to use a mix of protocols that are widely available. It’s easy to keep Modbus for simple reads/writes, use DNP3 where spotty links benefit from event buffers and time-stamps, and OPC UA where you want secure modeling across vendors.

    For sensor nets and edge telemetry, MQTT offers a lightweight publish/subscribe pattern suited to constrained, intermittent links.

    Selecting Hardware That Actually Lasts

    Environmental Protection

    Moisture, dust, and salt attack electronics. Match enclosures to the environment; NEMA 4X is common outdoors and in washdown or corrosive atmospheres. In classified areas, ensure the whole bill of materials (enclosure, fittings, devices) meets Class I, Div 2 rules.

    Power Resilience

    Power failure happens. Correctly size your UPS ride-through, pair with automatic transfer switches and generators following NFPA 110 guidance (design, testing, maintenance). Even when not legally required, adopting NFPA 110 conventions hardens recovery from grid events.

    Modularity

    Buy controllers with I/O headroom and communication-module expansion so you’re not having to start a whole new panel for a few new wells or more storage capacity.

    Software Platform Requirements

    • Each site’s control should be logically isolated even if infrastructure is shared. Role-based access ensures pumpers see controls, managers see summaries, and contractors see only what they need. OPC UA and modern SCADA platforms support certificate-based trust and user authorization patterns that align with this.
    • Push immediate alarms and safety logic to the edge so local automation carries the load when backhaul drops, a posture reinforced in OT security guidance.
    • Secure web/HMI views let supervisors acknowledge alarms, and techs fetch manuals and trends in the field—without poking holes around segmentation boundaries.

    Multi-Site Network Design

    Topology and links. Stars (each site back to HQ) are simple; meshes offer alternate paths over radio for resilience. Mix cellular primary with licensed radio failover; keep satellite as last-resort.

    Automatic failover. Let the comms layer switch paths without operator action. Prioritize alarm transport ahead of bulk history when bandwidth shrinks.

    Historian “store-and-forward.” Local buffers hold time-series data during outages and trickle it back when the link returns. Most modern historians and MQTT pipelines support this pattern out of the box; it’s a good antidote to compliance gaps from missing samples.

    Cloud vs. hybrid. Cloud deployment adds elasticity for analytics and storage, but pure cloud control adds risk. A hybrid model keeps critical functions on-prem while leveraging cloud to scale. That split is consistent with OT security references.

    Bandwidth hygiene. Use compression, report-by-exception, deadbands, and DNP3 event reporting so you’re not paying to move noise.

    Picking The Right Protocols

    • Modbus: ubiquitous, simple, minimal overhead; limited security features.
    • DNP3: event buffers, confirmations, secure authentication, time-sync; strong choice for unreliable links and compliance-friendly audit trails.
    • OPC UA: vendor-neutral information modeling with certificates for authentication, integrity, confidentiality; ideal for northbound IT/analytics.
    • MQTT: ultra-light pub/sub model that thrives on constrained links (battery sensors, remote skids), widely used across IoT and Oil & Gas applications.

    Compliance Integration (Make Audits Boring)

    Make Reporting Automatic

    Generate required forms directly from historian tags and events, templated per state (Texas RRC, OCC, OCD), with time-stamps and signatures handled electronically. 

    You’re aligning operations with the UIC Class II program while meeting local paperwork rules.

    Environmental Monitoring

    Fold groundwater, air, and spill detection into SCADA so alarms and trends live in the same pane of glass as injection metrics.

    Performance & Analytics

    Dashboards that matter. Surface injection efficiency, capacity headroom, equipment utilization, and energy burn. Use historian trends to justify capital or redistribute load.

    Predictive maintenance. Vibration and temperature patterns flag developing failures. Runtime counters move you from time-based to condition-based PMs; less wrench time, fewer surprises.

    Scheduling optimization. Blend reservoir response trends with trucking ETAs to maximize throughput without flirting with permit limits.

    Historical insight. Seasonal swings, gradual equipment decay, and energy cost patterns turn into targeted fixes and sensible budgets.

    What Good Looks Like in Practice

    • Operators get consistent screens across all sites and can triage alarms without hopping tools.
    • Maintenance sees condition trends and recommended actions, not cryptic tag floods.
    • Management tracks compliance posture, capacity headroom, and costs on one page.
    • Regulators receive clean, time-stamped reports aligned to their template—no manual re-entry.

    If you’re starting from scratch, build a thin slice first: two sites, standardized tags, historian with store-and-forward, segmented networks, and a minimal KPI dashboard. Then replicate.

    Dan Eaves

    Dan Eaves, PE, CSE

    Dan has been a registered Professional Engineer (PE) since 2016 and holds a Certified SCADA Engineer (CSE) credential. He joined PLC Construction & Engineering (PLC) in 2015 and has led the development and management of PLC’s Engineering Services Division. With over 15 years of hands-on experience in automation and control systems — including a decade focused on upstream and mid-stream oil & gas operations — Dan brings deep technical expertise and a results-driven mindset to every project.

    PLC Construction & Engineering (PLC) is a nationally recognized EPC company and contractor providing comprehensive, end-to-end project solutions. The company’s core services include Project Engineering & Design, SCADA, Automation & Control, Commissioning, Relief Systems and Flare Studies, Field Services, Construction, and Fabrication. PLC’s integrated approach allows clients to move seamlessly from concept to completion with in-house experts managing every phase of the process. By combining engineering precision, field expertise, and construction excellence, PLC delivers efficient, high-quality results that meet the complex demands of modern industrial and energy projects.

    Read more
    Piping Diagram

    From P&IDs to Panels: Specifying Control Panels and Passing FAT/SAT

    If you’ve ever watched a “simple” panel job turn into three weeks of scramble, you know the truth. The way we translate P&IDs into real, physical control panels makes or breaks commissioning. 

    Get the specification right and FAT/SAT feel like a formality. Miss a few details and you buy delays, field rework, and warranty heartburn.

    Here’s a practical, standards-anchored playbook so your panels ship right, install cleanly, and start up on schedule. From reading the P&IDs to closing out SAT.

    Understanding P&IDs and What They Don’t Tell You

    P&IDs are the backbone: they capture process flow, instruments, control loops, and protection functions you’ll marshal into a panel. 

    Use recognized symbol and identification standards so the whole team speaks the same language:

    • ISA-5.1 (Instrumentation Symbols & Identification).
    • ISO 14617-6 (graphical symbols for measurement/control).
    • PIP PIC001 practice for P&ID content and format.

    Read P&IDs methodically and extract a structured panel spec:

    • I/O & signals: per loop; type (AI/AO/DI/DO), ranges, isolation, power class, and any intrinsically safe barriers.
    • Safety integrity: which functions are SIS/SIF vs. BPCS, and the SIL target that will drive architecture and proof testing under IEC 61511 / ISA-84.
    • Communications: what must speak to what. EtherNet/IP, Modbus, OPC UA, and which links are safety-related vs. information only.
    • Environment & location: enclosure rating, temperature/humidity, corrosion exposure, and whether the panel or field devices sit in a hazardous (classified) location (e.g., Class I, Division 2 under NEC/NFPA 70/OSHA).

    Reality check: P&IDs rarely spell out alarm philosophy, historian tags, user roles, or cybersecurity boundaries; yet all of these affect the panel. 

    Close those gaps early using your site alarm standard (ISA-18.2 if you have it) and your OT security baseline (IEC 62443 / NIST SP 800-82).

    Specifying the Control Panel; Removing the Mystery

    1) Electrical and Safety Fundamentals

    • Applicable codes/standards: Design to UL 508A for industrial control panels (construction, component selection, SCCR, spacing/labels) and to NFPA 70 (NEC) for installation and hazardous-area rules. If you intend to ship a UL-labeled panel, say so explicitly in the spec.
    • Power architecture: feeder details, UPS/ride-through targets, heat load and cooling method, and fault/coordination assumptions that drive breaker and SCCR selections.
    • Arc-flash/LOTO hooks: provide nameplate data and working-clearance assumptions so the safety documentation and labels align with NEC/plant practice.

    2) Environmental and Enclosure Choices

    • Specify enclosure type rating and materials (e.g., 3R/4/4X) against salt/fog, washdown, or desert heat; define heater/AC setpoints and condensate routing. In hazardous locations, align construction with Class I, Division 2 expectations (equipment suitability, wiring methods, sealing).

    3) Networking and Cybersecurity by Design

    • Call out segmented networks (controls vs. corporate), managed switches, time sync, and remote-access methods. Reference IEC 62443 and NIST SP 800-82 so vendors document zones/conduits, authentication, and logging; not bolt them on later.

    4) HMI and Operator Experience

    • Define HMI size/brightness, glove/touch needs, language packs, and alarm colors/priorities to match your alarm philosophy. Good HMI rules save hours in SAT by avoiding “Where is that valve?” moments. Tie displays to tag names and cause-and-effect tables derived from the narrative.

    5) Documentation That is Actually Testable

    • Require: instrument index and I/O list, loop sheets, electrical schematics, network drawings, panel layout, bill of materials with certifications, software functional specification / control narrative, alarm rationalization tables, and FAT/SAT procedures. Quality documentation is the contract for acceptance.

    Functional Safety: Bake It In, Don’t Patch It Later

    If the panel carries any part of a SIS, treat those functions per IEC 61511 from day one:

    • Safety Requirements Specification (SRS).
    • Independence/separation from BPCS as required, diagnostics, bypass/override design, and proof-test intervals and methods captured in the test plan. 
    • Mapping P&ID cause-and-effect to SIFs early prevents last-minute rewires and retests.

    FAT: Make the Factory Your First Commissioning

    Why FAT matters: It’s cheaper to find mismatched wiring, wrong scaling, bad alarms, or flaky comms at the vendor’s bench than at your site. IEC 62381:2024 lays out the structure and checklists for FAT, FIT, SAT, and SIT. Use that backbone to avoid “interpretation debates.”

    Plan before you build:

    • Approve test procedures and acceptance criteria up front (I/O by I/O; sequences for start/stop/upset; comms failover; load/latency checks).
    • Define roles: who witnesses, who signs, who logs deviations/non-conformances.
    • Arrange the tooling: signal simulators, calibration gear, comms analyzers, and, for complex plants, a process simulator or emulation. (If you can’t simulate it, you can’t prove it.)

    Execute methodically:

    • I/O and loop checks: polarity, ranges, scaling, engineering units, clamps/limits, bumpless transfer, and fail-safe states.
    • Comms & integration: protocol verification (addressing, byte order, time-stamps), performance under load, and third-party skids integration.
    • Alarm tests: priorities and annunciation per your philosophy; standing-alarm rules; shelving/suppression behavior.
    • SIS proof points: for SIFs, demonstrate detection, logic, final element action, and trip times against SRS targets. Record what you prove and how often you must re-prove it.
    • Document everything: Log NCRs, corrective actions, and the as-tested configuration (firmware, IPs, logic versions). This package becomes the seed for SAT.

    SAT: Prove It in the Real World; Safely

    Between FAT and SAT, control drift happens (a device swap, a quick code fix). Lock versions, track MOC, and re-run targeted FAT steps if something changes.

    Prereqs worth confirming:

    • Power quality, grounding/bonding, and panel clearances match design; hazardous-area equipment and wiring meet NEC/OSHA expectations.
    • Network services (time sync, DHCP reservations, routes) actually exist on site, not just on the vendor’s bench.
    • Instruments are installed, calibrated, and ranged per the loop sheets.

    Run SAT in a deliberate order:

    1. Dry tests first (no live product): I/O point-to-point, permissives/interlocks proved with simulated signals.
    2. Cold commissioning: energize subsystems, check sequences without process risk.
    3. Live tests: exercise start/stop/abnormal scenarios with the process, record timings and loads, then compare to FAT baselines.
    4. Performance snapshots: capture response times, loop performance, and comms throughput as operating references for maintenance.

    Closeout with an operational turnover: as-builts, calibration certs, final programs/config backups, cause-and-effect, alarm philosophy, training records, and the signed FAT/SAT dossier.

    Common Trip-Wires and How to Step Around Them

    • Protocol quirks: Modbus register maps, byte order, and undocumented vendor “extensions” cause many delays. Specify and test protocol details during FAT; bring a sniffer.
    • Legacy surprises: Old PLCs/SCADA with limited connections or slow polling collapse under new loads. Identify limits early and throttle or upgrade.
    • Spec drift: small field changes stack into big test gaps. Control with formal change management tied to document versions.
    • Environment vs. build: panels that pass in a lab can fail in heat, dust, or salt. Size HVAC, coatings, and gasketing for reality, not brochures.
    • Hazardous area assumptions: labeling or wiring that doesn’t meet Class I, Div 2 or local code will halt SAT. Verify before shipment.

    A Minimal, High-Leverage Panel Spec

    • Standards: UL 508A build and label; NEC/NFPA 70 installation/hazardous location compliance.
    • Safety: IEC 61511 lifecycle for any SIF; SRS attached; proof-test intervals defined.
    • Docs: I/O index; loop sheets; schematics; panel GA; network drawings; bill of materials with certifications; control narrative; alarm philosophy; IEC 62381-aligned FAT/SAT plan.
    • Environment: enclosure rating (NEMA 4/4X/12), thermal design, corrosion/condensation mitigation; hazardous classification notes and wiring method.
    • Cyber: IEC 62443/NIST 800-82 references; zones/conduits; remote access/MFA; logging.

    Why This Works

    You’re aligning the design and test process with widely recognized guidance:

    • ISA-5.1 / ISO 14617 for drawings and symbols.
    • IEC 61511 / ISA-84 for safety.
    • IEC 62381 for FAT/SAT choreography.
    • UL 508A and NEC for how the panel is built and installed.
    • IEC 62443 / NIST 800-82 for security.

    That common language shortens meetings, sharpens acceptance criteria, and reduces surprises.

    Takeaways You Can Apply

    • Pick one pilot system and write the control narrative and FAT together; you’ll catch 80% of ambiguities before metal is bent.
    • Publish a one-page protocol sheet (addresses, registers, time sync, failover) to every vendor before FAT.
    • Add a site-readiness checklist to the SAT plan (power quality, grounding, network services, hazardous location verification).
    • Require a config snapshot (firmware/logic versions, IP plan) at FAT exit and at SAT entry—then diff them.
    Read more

    FEED vs. Detailed Design: How to De-Risk Your Gas Plant Build

    The Stakes and Why Front-End Choices Matter

    Gas plants are capital-intensive, multi-discipline beasts. Miss on scope or sequence and costs explode, schedules slip, and confidence fades. 

    In large capital programs broadly, reputable studies show chronic budget and schedule slippage, vast majorities of megaprojects run over. It is exactly why the front end has outsized leverage on outcomes.

    This guide clarifies when to use FEED (Front End Engineering Design) versus jumping straight to detailed design, and how that choice affects risk, cost accuracy, procurement timing, and delivery.

    FEED vs. Detailed Design: What’s the Real Difference?

    What FEED Actually Does

    Think of FEED as the bridge from concept to buildable intent. It locks the process basis and key design criteria, producing PFDs/P&IDs, plot plans, preliminary equipment specs, and the safety backbone. HAZOP and SIS/SIL planning per IEC standards, for example.

    The payoffs are tighter estimates and fewer surprises. In the AACE estimate-class framework, moving from conceptual (Class 5/4) toward Class 3 typically improves accuracy to roughly −10%/−20% to +10%/+30% depending on complexity. It is far better than the ±50% conceptual range often cited for early studies.

    On cost, FEED commonly falls in the ~2–3% of TIC range (some programs cite ~3–5% depending on depth and complexity), but that spend underwrites sharper scope, procurement strategy, and construction planning.

    Safety and operability analysis belong here. Use IEC 61882 to structure HAZOPs and IEC 61511 to frame SIS lifecycle/SIL determination.

    What Detailed Design Delivers

    Detailed design transforms FEED intent into construction-ready drawings, isometrics, data sheets, cable schedules, control narratives, and procurement packages across mechanical, electrical, I&E, civil/structural. At this point, estimate precision typically tightens again (toward Class 2/1 bands) with narrower ranges suitable for bids and Final Investment Decisions.

    The distinction in plain language: FEED defines the right plant; detailed design defines every bolt of the right plant.

    Why Front-End Rigor Pays (Risk, Cost, Schedule)

    1) Cost Accuracy You Can Defend

    FEED narrows uncertainty from conceptual swings (±50%) toward Class-3-like ranges (often in the ~±15–30% envelope for process industries), enabling credible budgets, contracting strategy, and financing.

    Independent and vendor literature report that robust FEED/front-end loading correlates with lower total installed cost and shorter execution. Some benchmarks cite material cost reductions and schedule improvements when FEED is thorough.

    2) Fewer Technical Surprises

    FEED is where you validate process simulations, unit operations, and operability/maintainability. Run HAZOP; assign SIL to SIFs; and stress-test tie-ins.

    Doing it here prevents costlier changes later and anchors mandatory safety/protection requirements for detailed design.

    3) Schedule You Can Actually Hit

    A complete FEED lets you start procurement in parallel with detailed design. Critical when long-lead packages (e.g., large compressors, major electrical gear) run 20–60 weeks or more, with some compression equipment out a year+ in tight markets. Early identification and pre-bid/vendor engagement protect the critical path.

    For LNG and large gas processing, FEED itself can take 12–18 months, but that time produces a package that de-risks EPC and informs long-lead buys.

    Gas-Plant Reality Check: The Risk Landscape

    • Technical/Process: complex separations, acid gas removal, sulfur recovery, and high-consequence safety envelopes—best handled with standards-driven HAZOP/SIS governance.
    • Commercial: inflation, supply-chain volatility, and scarce specialty resources. Sector-wide analyses still see 30–45% average budget/schedule variance on major programs without better controls/visibility.
    • Regulatory/ESG: tightening emissions and permitting expectations add steps you want planned, not discovered.
    • Operational: 20–30-year life cycles demand flexibility for feed changes and future debottlenecking.

    How FEED Protects Your Project

    Cost discipline early. Use FEED to standardize equipment, simplify process trains, and remove bespoke one-offs. Lock an AACE-aligned basis of estimate and contingency logic; socialize it with financiers and partners to avoid late-stage resets.

    Safety first, on paper. Complete HAZOP and define SIS lifecycle/SIL targets before detailed design. Treat the outputs as design requirements, not advice.

    Procurement strategy early. Identify long-lead items during FEED; pre-qualify vendors and launch RFPs on the first safe opportunity. Many MEP/electrical packages (switchgear, AHUs, large valves) now see 20–60-week windows; large compression skids may extend 12+ months.

    Parallelize smartly. With process requirements frozen and key specs set, detailed design can progress while long-lead orders and early works start—shortening your critical path.

    When to Move from FEED to Detailed Design

    Green lights typically require:

    • Technical maturity: simulations closed, FEED-level P&IDs/plot plan, HAZOP actions addressed, preliminary 3D/constructability passes done.
    • Commercial readiness: budget approved, funding plan in place, contracting model set, long-lead procurement strategy defined.
    • Permitting/ESG: material approvals on track to avoid EPC stalls.
    • Risk posture: if you must accelerate, quantify what’s “at risk” (and cap it). Sector analyses warn that under-cooked front ends are a common root cause of overrun/overrun.

    Managing Gas-Plant Risk

    1. Own a living risk register from FEED onward, covering technical/commercial/schedule/regulatory line items with owners and triggers.
    2. Favor proven tech unless the business case justifies pilot/prototype risk; if you must push tech, secure vendor guarantees and performance bonds.
    3. Plan compliance in FEED—early agency engagement, environmental baseline work, and submissions sequenced to your long-lead timeline.
    4. Build real contingency: technical alternates, schedule recovery options, and cost mitigation actions you can actually execute.

    Making the Choice

    • Use thorough FEED for high-complexity, first-of-a-kind, brownfield tie-ins, constrained sites, or tight safety envelopes.
    • Consider acceleration only when market timing justifies it and you can quantify the added risk (and carry the contingency).
    • Don’t skip FEED to “save” 2–3%—front-end investment routinely saves multiples downstream via fewer changes, cleaner procurement, and faster commissioning.
    • Match to capability: experienced owner’s teams may compress phases; others should buy rigor with experienced FEED partners.

    A Simple Decision Checklist

    • Scope clarity: Process basis frozen? Battery limits clear?
    • Safety: HAZOP complete (key actions closed), SIS/SIL targets set per IEC 61511?
    • Estimate maturity: AACE-aligned class with documented assumptions/contingency?
    • Procurement: Long-lead list finalized; RFPs/tenders staged; vendor shortlist agreed?
    • Schedule logic: FEED→detailed design overlap defined; early works identified; critical path driven by long-lead reality (not hope)?
    • Permits/ESG: filings sequenced to avoid EPC stalls?
    • Change control: frozen-line philosophy and governance in place?
    1. Adopt a narrative template mapped to ISA-106 states plus ISA-18.2 alarm hooks; pilot it on one complex unit.
    2. Publish an alarm philosophy one-pager (priorities, KPIs, standing-alarm rules) and socialize it at the console.
    3. Stand up a role-based training index tied to your OQ program (API RP 1161/PHMSA FAQs) so every trainee knows the modules to complete before CSU.
    Read more
    Using a Training Manual for Guidance

    Control Narratives And Training Manuals: Documentation That Speeds Startup

    The Quiet Throughput Killer and the Fix You Control

    Every extra day a facility limps through startup costs real money; missed production, overtime, rentals, and reputational drag. The countermeasure isn’t exotic: strong documentation.

    Treat control narratives and training manuals as core deliverables. Plan, standardize, and tie to how the plant actually behaves. Startups run cleaner and commissioning teams spend less time firefighting and more time verifying.

    Across the process industries, guidance from the Construction Industry Institute (CII), ISA, and regulators consistently ties clearer procedures and commissioning discipline to better outcomes in startup and early operations. 

    CII’s commissioning/startup research highlights critical success factors that include robust documentation and defined procedures, not just “as-built” drawings.

    Why Documentation Accelerates Commissioning

    Documentation isn’t overhead. It’s the operating system for startup. With complete, searchable, and consistent narratives/manuals:

    • Field teams diagnose faster because the intended control behavior is explicit—not buried in tribal knowledge.
    • Work can run in parallel: operators train while construction wraps, maintenance stages spares against documented BOMs, and supervisors finalize procedures grounded in the same logic the control system uses.
    • Commissioning and alarm handling lean on recognized practices, reducing noise and chasing only “real” alarms; like ISA-18.2 alarm lifecycle.

    CII’s commissioning/startup body of work calls out procedure quality, turnover packages, and disciplined CSU planning as critical success factors. The things that statistically show up on projects that meet schedule and performance targets.

    Control Narratives: Make System Behavior Unambiguous

    A control narrative explains what the automation does, not just what a person should do.

    It translates P&IDs and control philosophies into plain-language sequences for normal, abnormal, and shutdown states. These include cause-and-effect logic, permissives, interlocks, timing, and alarms. That clarity is priceless during first-fire and upset testing.

    Anchor your format to standards so every system reads the same way:

    • Alarm behavior and responses consistent with ISA-18.2 (alarm philosophy, rationalization, KPIs).
    • Procedural automation patterns per ISA-106 (models, styles, lifecycle for automating procedures in continuous processes).
    • Safety functions linked to the IEC 61511 lifecycle (SIS/SIL targets captured in Safety Requirements Specifications, then reflected in the narrative and logic).

    What Good Looks Like: Fast on the Eyes, Useful in the Field

    • Structure: Overview → modes/states → sequences (start/normal/stop/upset) → interlocks & permissives → alarms & operator actions → fail-safe behavior.
    • Language: Concrete, stepwise conditions (“If suction pressure < X for ≥ Y s, then close Z; if not cleared in ≤ T s, raise Alarm A with priority B”). Avoid vague qualifiers.
    • Cross-links: P&IDs, loop sheets, cause-and-effects, alarm philosophy, and HMI screenshots are interlinked so techs pivot in one click.
    • Version control: Revisions tied to MOC; field redlines reconcile into the master narrative before turnover.

    Training Manuals: Turning Book Learning Into Safe, Fast Competence

    Great manuals shorten time-to-competency by pairing tasks with the why behind them. Especially where safety and reliability depend on correct first actions.

    Build on sector references and regulatory frameworks:

    • For pipeline and midstream operations, API RP 1161 lays out Operator Qualification (OQ) program guidance; PHMSA provides OQ FAQs that clarify expectations. Use these to shape job task analyses, qualification methods, and refresher cycles.
    • Align alarm/abnormal response training with your ISA-18.2 alarm philosophy so operators learn the system they’ll actually see.

    Manual Design That Sticks Under Pressure

    • Organized by job role and scenario, not department. Operators get state-based playbooks; maintenance gets condition-based checks; supervisors get shift-change and escalation flows.
    • Multiple modalities: diagrams and flows for visual learners; narrations or brief videos for auditory learners; walkthrough/simulation drills for kinesthetic learners.
    • Decision aids: concise fault trees and “first five minutes” cards for high-stress events.
    • Competency gates: short checks at the end of each module tied to OQ or site standards, with remedial loops if a trainee struggles on a step that’s safety-critical.

    How Documentation Speeds Startup

    1. Faster diagnosis. When commissioning trips a shutdown, the narrative points straight to the interlock logic and intended operator response; teams fix causes, not symptoms.
    2. Parallelism. While I&E closes punch items, ops can train against the same logic the PLC will run; maintenance can pre-position spares from approved data sheets.
    3. Cleaner alarms. Documented alarm philosophy (ISA-18.2) trims nuisance alarms, focuses attention, and reduces alarm floods during startup transients.
    4. Safer handovers. Narratives and manuals become the backbone of CSU turnover packages highlighted in CII commissioning/startup guidance.

    Control Narrative Best Practices

    1) Standardize the skeleton. Use one template across units: states, transitions, timing, permissives, interlocks, alarms, failsafes, and manual interventions.

    2) Write for the reader. Keep syntax consistent and testable. Every condition is measurable; every response has a time base and priority.

    3) Tie to safety from day one. When a HAZOP or LOPA assigns a SIF and SIL, update the narrative and tag references so the SIS logic and BPCS are coherent.

    4) Make it navigable. Hyperlink P&IDs, loop sheets, alarm rationalization tables, and HMI mockups; build the same links into your CMMS and historian so techs can jump from an alarm to the narrative and then to the work order.

    5) Control the versions. No “mystery PDFs.” Check in/out through your document control; link MOC numbers to each revision.

    Training Manual Excellence Built for Real Plants

    1) Start from tasks. Derive modules from a role’s critical tasks (as OQ/PHMSA frameworks expect), then teach why the task and the system behavior matter.

    2) Simulate realities. Drills on start/stop, loss of utility, upset recovery, and alarm floods build true confidence.

    3) Keep it plain. Define site-specific terms. Side-bars for “common pitfalls” and “don’t do this” moments.

    4) Measure and adapt. Put quick checks at the end of every module and trend time-to-competency; close gaps with micro-lessons, not just longer manuals.

    Implementation Playbook

    Phase 1: Foundation

    • Draft your alarm philosophy (ISA-18.2) and narrative template (with ISA-106/IEC-61511 hooks).
    • Inventory systems needing narratives; prioritize safety-critical and high-complexity units first.

    Phase 2: Write and Wire

    • Write narratives alongside control logic development; cross-link tags and sequences.
    • Build training modules from those same narratives and your OQ task list (API RP 1161).

    Phase 3: Prove It Before Startup

    • Dry-run procedures in a FAT/SAT context; test alarm rates against philosophy targets; fix gaps in docs and HMI language. This aligns with CII’s CSU best-practice emphasis on readiness and defined turnover.

    Phase 4: Turnover & Sustain

    • Deliver a navigable package: narratives, alarm philosophy, HMI guide, data sheets, P&IDs, and training modules—version-controlled and searchable.
    • Put reviews on the calendar: post-startup 30/60/90-day edits, then quarterly light updates and annual full sweeps.

    KPIs That Show the Payoff

    • Commissioning first-pass yield (tests accepted on first try).
    • Alarm health (ISA-18.2 KPIs: standing alarms, alarms/hour/operator at steady state, top offenders).
    • Time-to-competency for new roles (aligned with OQ expectations).
    • Post-startup change rate (number of logic/document changes in first 90 days).
    • Mean time to diagnose top 10 faults (trend down as narratives improve).

    Common Pitfalls and Quick Fixes

    • Vague logic. Replace “when pressure is high” with thresholds, deadbands, and timers.
    • Document drift. Tie every code change to a document update via MOC.
    • Alarm floods. Rationalize against ISA-18.2; demote, suppress (safely), or eliminate chaff.
    • Training that’s “read-only.” Add scenario drills and short, role-based refreshers keyed to recent incidents and bad-actor alarms.

    Your First Three Moves

    1. Adopt a narrative template mapped to ISA-106 states plus ISA-18.2 alarm hooks; pilot it on one complex unit.
    2. Publish an alarm philosophy one-pager (priorities, KPIs, standing-alarm rules) and socialize it at the console.
    3. Stand up a role-based training index tied to your OQ program (API RP 1161/PHMSA FAQs) so every trainee knows the modules to complete before CSU.
    Read more
    Industrial Compressor Station

    Compressor Station Optimization: Advanced Control Strategies That Cut Downtime

    Why Compressor Failures Hurt So Much

    When a station trips, the losses stack fast. Lost throughput, missed nominations, penalty risk, and crews scrambling. Across heavy industry, recent surveys peg the typical cost of unplanned downtime in the six figures per hour (and higher for some operations).

    This is why “run-to-fail” is no longer a plan; if it ever was.

    Modern control rooms don’t live on clipboards. They run real-time analytics, predictive models, and alarm logic tuned to catch weak signals early. 

    Done well, predictive maintenance programs have been shown to cut downtime by up to ~50% and reduce maintenance costs by roughly 10–40%, depending on asset class and maturity.

    Optimization isn’t a slogan; it’s an operating posture.

    • Regulators push for cleaner, safer operations.
    • Customers want reliability.
    • Investors watch margins.

    The only way to satisfy all three is to squeeze more availability and efficiency from every compressor you own; without gambling on reliability.

    What “Optimization” Really Means

    Strip the jargon and you’re left with three outcomes: higher availability, lower energy per unit moved, and fewer surprises. World-class operations measure it. 

    As a rough compass, OEE benchmarks often cite 85% as “world-class” while many plants live closer to 60–70%, leaving obvious headroom. Your mix will vary, but the message is the same: there’s room to improve.

    Yesterday’s “optimize” meant looking at last month’s report and scheduling PM on a calendar. Today it means sensor networks, models that learn normal behavior, and controls that adjust in real time. It’s the difference between driving by the rear-view mirror and using live GPS.

    Expectations You Can Defend

    • Energy: Real-time optimization and set-point management in gas networks have delivered single- to double-digit reductions in compression energy in studies of dynamic operation (examples report ~5–8% savings in specific networks, with larger gains in some cases).
    • Maintenance: Data-driven programs consistently show double-digit cost reductions and large cuts in unplanned downtime when compared with purely time-based maintenance.
    • Reliability: Standard OT security and architecture (segmentation, least-privilege remote access) materially lowers the chance that cyber issues trigger operational incidents.

    Control Strategies That Actually Move the Needle

    1) See Problems Before They’re Problems

    Blanket monitoring isn’t the goal; discriminating monitoring is. 

    Vibration spectra, thermography, lube oil analysis, and process signals together spot the weak signals of bearing wear, misalignment, cooling issues, and electrical faults; months before a failure

    Machine-learning models then learn each machine’s “normal” and flag drift, so your team schedules work on your terms, not the asset’s. Studies across industries tie PdM to up to ~50% downtime reduction and 10–40% maintenance-cost cuts when it’s implemented well.

    Integrate this with your CMMS so work is triggered by condition: don’t change oil at 2,000 hours if analysis says it’s healthy; don’t wait if contamination spikes at 1,500. 

    Over time, your historian reveals true useful life:

    • “that bearing averages 18 months”
    • “those seals last ~5,000 cycles”

    So spares and windows are planned, not panicked.

    2) Load and Pressure Strategy, Not Guesswork

    Why run five machines at 70% when three at their sweet spot will do? Use smart load sharing to keep units near their best efficiency islands, and rotate runners to spread wear.

    Move from fixed discharge setpoints to dynamic pressure control that considers downstream demand, compressor maps, and fuel or power price. Delivering the same throughput with less energy and less equipment stress. 

    Pipeline studies and operations research continue to show energy savings by optimizing compression under transient conditions.

    Forecasts matter here. Blend weather, supply nominations, and historical profiles so the station ramps before the morning surge, not during it, shaving peak-period energy without risking service.

    3) Dashboards That Help Humans

    Operators don’t need ten screens of tiny numbers. They need compression efficiency, energy per MSCF, equipment health, and throughput; with drill-downs one click away.

    Alerts should be context-aware to kill alarm fatigue: group symptoms, route by severity, and escalate with evidence.

    4) Design for Fast Recovery

    Failures still happen. The difference between a stumble and a crisis is hot-standby designs, documented recovery (with tooling staged), and people who’ve rehearsed it.

    If your cost of downtime is measured in six figures per hour, it doesn’t take many averted hours to justify selective redundancy. 

    Industry surveys routinely show high downtime cost ranges, which is why resilience pays for itself.

    A Playbook To Making It Real

    Phase 1: Instrument and Baseline

    Start with health and energy instrumentation that feeds a historian. Establish KPIs: availability, trips per 1,000 hours, energy per throughput, mean time between failures.

    Phase 2: Predict and Prioritize

    Add analytics that spot early-warning patterns. Tie actions to the CMMS. Use the first quarter to validate “find-fix” loops and tune alarm logic.

    Phase 3: Automate Controls

    Introduce load-sharing and dynamic pressure control with clear guardrails. Pilot on one station, then templatize.

    Phase 4: Standardize and Scale

    Lock down naming, templates, spares, and playbooks so operators can cover multiple stations without relearning the HMI every time.

    Security and Compliance

    Industrial cybersecurity isn’t optional. 

    • Use zones and conduits to segment networks
    • Add a DMZ for business data
    • Enforce least-privilege, MFA-protected remote access
    • Instrument for detection.

    The current NIST SP 800-82 Rev. 3 and the ISA/IEC 62443 series are the go-to playbooks for ICS/OT architecture and controls. Build them and you’ll satisfy both risk and audit.

    What to Measure And Prove

    • Availability (target ≥95% for critical lines) and MTBF by unit.
    • Energy per unit throughput and compressor efficiency trends; verify savings against weather and demand to avoid “phantom” gains. Academic and OR literature on gas networks documents measurable energy cuts from dynamic compressor optimization.
    • Maintenance mix: % reactive vs. planned vs. condition-based; trend toward more CBM.
    • Alarm quality: rate, deduplication, time-to-ack; aim for fewer, better alerts.

    Budget, Timing, and ROI

    You don’t have to big-bang this. Many teams see early wins inside a quarter once monitoring is live and work orders are tied to conditions.

    Industry reports and surveys consistently show double-digit O&M savings and meaningful downtime reductions from predictive programs. Some organizations report paybacks inside 6–18 months; asset class and starting point drive the spread.

    On the capital side, avoid gold-plating redundancy. Protect the true bottlenecks first: drivers, key auxiliaries, controls infrastructure. If your outage cost sits in six figures per hour, even a modest reduction in event frequency or duration closes the loop quickly.

    Putting the Pieces Together

    • Health Monitoring: vibration, temperature, lube, and process KPIs wired to a historian; anomaly detection tuned per unit.
    • Load/Pressure Control: automate load sharing; replace fixed setpoints with demand-aware pressure targets. Validate energy savings against baseline.
    • Human-Centered HMI: KPI-first displays; alarm suppression and routing that cut noise.
    • Playbooks & Spares: standard recovery steps; staged tools; minimum spares for known failure modes.
    • Security: segmented networks, MFA remote access, logging/monitoring that operations can actually use.

    Bottom Line

    Advanced control isn’t a luxury, it’s how competitive stations run. Predictive maintenance, intelligent load management, better HMIs, and solid security raise availability, cut energy, and reduce surprises.

    Start with an honest baseline, pilot where the payoff is obvious, and scale what works.

    If you want help prioritizing the first pilot, begin with the station showing the worst energy per MSCF and highest alarm rate. That combination usually hides the fastest wins.

    Read more