home / skills / pluginagentmarketplace / custom-plugin-mongodb / mongodb-integration
This skill helps you master MongoDB integration across Node.js, Python, and Java apps, covering connections, transactions, error handling, and best practices.
npx playbooks add skill pluginagentmarketplace/custom-plugin-mongodb --skill mongodb-integrationReview the files below or copy the command above to add this skill to your agents.
---
name: mongodb-app-development
version: "2.1.0"
description: Master MongoDB integration in applications with Node.js, Python, and Java drivers. Learn connections, transactions, error handling, and best practices. Use when building applications with MongoDB.
sasmp_version: "1.3.0"
bonded_agent: 07-mongodb-application-development
bond_type: PRIMARY_BOND
# Production-Grade Skill Configuration
capabilities:
- driver-integration
- connection-management
- error-handling
- transaction-implementation
- change-streams
input_validation:
required_context:
- programming_language
- operation_type
optional_context:
- framework
- async_requirement
- connection_pool_size
output_format:
code_implementation: string
error_handling: string
connection_config: object
best_practices: array
error_handling:
common_errors:
- code: APP001
condition: "Connection pool exhausted"
recovery: "Increase maxPoolSize, check for connection leaks"
- code: APP002
condition: "Network timeout"
recovery: "Increase socketTimeoutMS, check network stability"
- code: APP003
condition: "Driver version incompatibility"
recovery: "Update driver to latest stable version"
prerequisites:
mongodb_version: "4.0+"
required_knowledge:
- chosen-language-basics
- mongodb-crud
driver_requirements:
- "Official MongoDB driver for target language"
testing:
unit_test_template: |
// Integration test
beforeAll(async () => {
await client.connect()
})
afterAll(async () => {
await client.close()
})
it('should perform CRUD operations', async () => {
const result = await collection.insertOne({ test: true })
expect(result.insertedId).toBeDefined()
})
---
# MongoDB Application Development
Master MongoDB integration in production applications.
## Quick Start
### Node.js Driver Setup
```javascript
const { MongoClient } = require('mongodb');
const client = new MongoClient('mongodb://localhost:27017', {
useNewUrlParser: true,
useUnifiedTopology: true,
maxPoolSize: 10
});
async function main() {
try {
await client.connect();
const db = client.db('myapp');
const collection = db.collection('users');
// Use collection
} finally {
await client.close();
}
}
main().catch(console.error);
```
### Connection Pooling
```javascript
// Configure connection pool
const client = new MongoClient(
'mongodb://localhost:27017',
{
maxPoolSize: 10, // Maximum connections
minPoolSize: 5, // Minimum connections
maxIdleTimeMS: 45000, // Idle timeout
serverSelectionTimeoutMS: 5000
}
);
```
### Transactions
```javascript
const session = client.startSession();
try {
await session.withTransaction(async () => {
// Multiple operations atomic together
await users.insertOne({ name: 'John' }, { session });
await accounts.insertOne({ userId: '...', balance: 100 }, { session });
// If any fails, all roll back
});
} catch (error) {
console.error('Transaction failed:', error);
} finally {
await session.endSession();
}
```
### Error Handling
```javascript
try {
await collection.insertOne({ email: '[email protected]' });
} catch (error) {
if (error.code === 11000) {
console.log('Duplicate key error');
} else if (error.name === 'MongoNetworkError') {
console.log('Network error - retry');
} else if (error.name === 'MongoWriteConcernError') {
console.log('Write concern failed');
} else {
console.log('Other error:', error);
}
}
```
### Bulk Operations
```javascript
const bulk = collection.initializeUnorderedBulkOp();
bulk.insert({ name: 'John', age: 30 });
bulk.insert({ name: 'Jane', age: 28 });
bulk.find({ age: { $lt: 25 } }).update({ $set: { status: 'young' } });
bulk.find({ name: 'old_user' }).remove();
await bulk.execute();
```
## Change Streams
```javascript
// Watch collection for changes
const changeStream = collection.watch();
changeStream.on('change', (change) => {
console.log('Change:', change.operationType);
console.log('Document:', change.fullDocument);
});
// With pipeline filtering
const pipeline = [
{ $match: { 'operationType': 'insert' } },
{ $match: { 'fullDocument.status': 'active' } }
];
const filteredStream = collection.watch(pipeline);
```
## Python Integration
```python
from pymongo import MongoClient
from pymongo.errors import DuplicateKeyError
import asyncio
# Synchronous
client = MongoClient('mongodb://localhost:27017')
db = client['myapp']
users = db['users']
# CRUD
user = users.insert_one({'name': 'John', 'email': '[email protected]'})
found = users.find_one({'_id': user.inserted_id})
users.update_one({'_id': user.inserted_id}, {'$set': {'age': 30}})
users.delete_one({'_id': user.inserted_id})
# Transactions
def transfer_funds(from_user_id, to_user_id, amount):
with client.start_session() as session:
with session.start_transaction():
users.update_one(
{'_id': from_user_id},
{'$inc': {'balance': -amount}},
session=session
)
users.update_one(
{'_id': to_user_id},
{'$inc': {'balance': amount}},
session=session
)
# Error handling
try:
users.insert_one({'email': '[email protected]'})
except DuplicateKeyError:
print('Email already exists')
```
## Java Integration
```java
import com.mongodb.client.MongoClient;
import com.mongodb.client.MongoClients;
import com.mongodb.client.MongoDatabase;
import com.mongodb.client.MongoCollection;
import org.bson.Document;
public class MongoDB {
public static void main(String[] args) {
// Connect
MongoClient mongoClient = MongoClients.create("mongodb://localhost:27017");
MongoDatabase database = mongoClient.getDatabase("myapp");
MongoCollection<Document> users = database.getCollection("users");
// Insert
Document user = new Document("name", "John")
.append("email", "[email protected]")
.append("age", 30);
users.insertOne(user);
// Find
Document found = users.find(new Document("email", "[email protected]")).first();
// Update
users.updateOne(
new Document("email", "[email protected]"),
new Document("$set", new Document("age", 31))
);
// Delete
users.deleteOne(new Document("email", "[email protected]"));
// Close
mongoClient.close();
}
}
```
## Performance Patterns
```javascript
// Pagination
async function paginate(collection, page = 1, limit = 10) {
const skip = (page - 1) * limit;
return collection
.find({})
.skip(skip)
.limit(limit)
.toArray();
}
// Batch processing
async function processBatch(collection, batchSize = 1000) {
const cursor = collection.find({});
let batch = [];
for (const doc of await cursor.toArray()) {
batch.push(doc);
if (batch.length === batchSize) {
await processBatch(batch);
batch = [];
}
}
if (batch.length > 0) {
await processBatch(batch);
}
}
// Lazy loading
class LazyLoader {
constructor(collection, query) {
this.cursor = collection.find(query).batchSize(100);
}
async next() {
return this.cursor.hasNext() ? this.cursor.next() : null;
}
}
```
## Best Practices
✅ Always close connections properly
✅ Use connection pooling
✅ Handle errors appropriately
✅ Use transactions for multi-document operations
✅ Validate input data
✅ Use appropriate write concerns
✅ Monitor connection status
✅ Implement retry logic
✅ Use indexes for queries
✅ Log important operations
✅ Test error scenarios
✅ Use async/await for Node.js
This skill teaches practical MongoDB integration for application development with Node.js, Python, and Java drivers. It focuses on reliable connections, transactions, error handling, bulk ops, change streams, and performance patterns. Followable code snippets and patterns help you build robust, production-ready database interactions.
It shows how to configure clients, manage connection pools, and open/close sessions correctly. The skill demonstrates transactions for atomic multi-document operations, bulk operations for throughput, and change streams for real-time updates. Error handling patterns and performance techniques like pagination, batching, and lazy loading are provided for each supported driver.
Do I need transactions for every write?
No. Use transactions when multiple documents or collections must change atomically; single-document operations are atomic by default.
How do I handle duplicate key errors?
Catch DuplicateKeyError (Python) or error code 11000 (Node/Java) and implement a conflict resolution or user-facing message.
When should I use change streams?
Use change streams when you need low-latency reaction to data changes for real-time features, auditing, or syncing services; filter with a pipeline to reduce noise.
How to choose pool sizes?
Base max/min pool size on expected concurrent operations, connection limits of your deployment, and latency. Start conservative, monitor metrics, and adjust.