home / skills / athola / claude-night-market / architecture-paradigm-space-based
/plugins/archetypes/skills/architecture-paradigm-space-based
This skill helps optimize high-traffic stateful workloads by applying space-based data grid paradigms for elastic, linear in-memory scalability.
npx playbooks add skill athola/claude-night-market --skill architecture-paradigm-space-basedReview the files below or copy the command above to add this skill to your agents.
---
name: architecture-paradigm-space-based
description: 'Data-grid architecture for high-traffic stateful workloads with linear
scalability.
space-based, data grid, in-memory, linear scaling, high traffic
Use when: traffic overwhelms database nodes or linear scalability needed
DO NOT use when: data does not fit in memory or simpler caching suffices.'
category: architectural-pattern
tags:
- architecture
- space-based
- data-grid
- scalability
- in-memory
- stateful
dependencies: []
tools:
- data-grid-platform
- replication-manager
- load-tester
usage_patterns:
- paradigm-implementation
- high-traffic-workloads
- linear-scalability
complexity: high
estimated_tokens: 800
---
# The Space-Based Architecture Paradigm
## When To Use
- High-traffic applications needing elastic scalability
- Systems requiring in-memory data grids
## When NOT To Use
- Low-traffic applications where distributed caching is overkill
- Systems with strong consistency requirements over availability
## When to Employ This Paradigm
- When traffic or state volume overwhelms a single database node.
- When latency requirements demand in-memory data grids located close to processing units.
- When linear scalability is required, achieved by partitioning workloads across many identical, self-sufficient units.
## Adoption Steps
1. **Partition Workloads**: Divide traffic and data into processing units, each backed by a replicated data cache.
2. **Design the Data Grid**: Select the appropriate caching technology, replication strategy (synchronous vs. asynchronous), and data eviction policies.
3. **Coordinate Persistence**: Implement a write-through or write-behind strategy to a durable data store, including reconciliation processes.
4. **Implement Failover Handling**: Design a mechanism for leader election or heartbeats to validate recovery from node loss without data loss.
5. **Validate Scalability**: Conduct load and chaos testing to confirm the system's elasticity and self-healing capabilities.
## Key Deliverables
- An Architecture Decision Record (ADR) detailing the chosen grid technology, partitioning scheme, and durability strategy.
- Runbooks for scaling processing units and for recovering from "split-brain" scenarios.
- A monitoring suite to track cache hit rates, replication lag, and failover events.
## Risks & Mitigations
- **Eventual Consistency Issues**:
- **Mitigation**: Formally document data-freshness Service Level Agreements (SLAs) and implement compensation logic for data that is not immediately consistent.
- **Operational Complexity**:
- **Mitigation**: The orchestration of a data grid requires mature automation. Invest in production-grade tooling and automation early in the process.
- **Cost**:
- **Mitigation**: In-memory grids can be resource-intensive. Implement aggressive monitoring of utilization and auto-scaling policies to manage costs effectively.
## Troubleshooting
### Common Issues
**Command not found**
Ensure all dependencies are installed and in PATH
**Permission errors**
Check file permissions and run with appropriate privileges
**Unexpected behavior**
Enable verbose logging with `--verbose` flag
This skill describes the space-based (data-grid) architecture for high-traffic stateful workloads that require linear scalability. It explains how to partition state and processing into many identical, in-memory units so systems scale elastically and maintain low latency under heavy load. The guidance targets engineers deciding when and how to adopt an in-memory data grid for production services.
The approach partitions traffic and state across multiple processing units, each backed by a replicated in-memory cache so work is handled locally and in parallel. Persistence is coordinated via write-through or write-behind strategies to a durable store, while leader election and heartbeat-based failover handle node loss. Monitoring and chaos testing validate linear scaling and self-healing behavior.
When should I NOT use space-based architecture?
Avoid it if your working set does not fit in memory, if a simpler caching layer solves the problem, or if strict strong consistency is mandatory.
How do I ensure durability with an in-memory grid?
Use write-through or write-behind to a durable datastore and implement reconciliation processes plus monitoring for replication lag and failed writes.