Compare commits
2 Commits
71b2855d01
...
try-fix
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
4378146d0c | ||
|
|
b26002eff9 |
34
.env.example
34
.env.example
@@ -1,34 +0,0 @@
|
|||||||
# ===============================================
|
|
||||||
# DNSRecon Environment Variables
|
|
||||||
# ===============================================
|
|
||||||
# Copy this file to .env and fill in your values.
|
|
||||||
|
|
||||||
# --- API Keys ---
|
|
||||||
# Add your Shodan API key for the Shodan provider to be enabled.
|
|
||||||
SHODAN_API_KEY=
|
|
||||||
|
|
||||||
# --- Flask & Session Settings ---
|
|
||||||
# A strong, random secret key is crucial for session security.
|
|
||||||
FLASK_SECRET_KEY=your-very-secret-and-random-key-here
|
|
||||||
FLASK_HOST=127.0.0.1
|
|
||||||
FLASK_PORT=5000
|
|
||||||
FLASK_DEBUG=True
|
|
||||||
# How long a user's session in the browser lasts (in hours).
|
|
||||||
FLASK_PERMANENT_SESSION_LIFETIME_HOURS=2
|
|
||||||
# How long inactive scanner data is stored in Redis (in minutes).
|
|
||||||
SESSION_TIMEOUT_MINUTES=60
|
|
||||||
|
|
||||||
|
|
||||||
# --- Application Core Settings ---
|
|
||||||
# The default number of levels to recurse when scanning.
|
|
||||||
DEFAULT_RECURSION_DEPTH=2
|
|
||||||
# Default timeout for provider API requests in seconds.
|
|
||||||
DEFAULT_TIMEOUT=30
|
|
||||||
# The number of concurrent provider requests to make.
|
|
||||||
MAX_CONCURRENT_REQUESTS=5
|
|
||||||
# The number of results from a provider that triggers the "large entity" grouping.
|
|
||||||
LARGE_ENTITY_THRESHOLD=100
|
|
||||||
# The number of times to retry a target if a provider fails.
|
|
||||||
MAX_RETRIES_PER_TARGET=8
|
|
||||||
# How long cached provider responses are stored (in hours).
|
|
||||||
CACHE_EXPIRY_HOURS=12
|
|
||||||
1
.gitignore
vendored
1
.gitignore
vendored
@@ -169,3 +169,4 @@ cython_debug/
|
|||||||
#.idea/
|
#.idea/
|
||||||
|
|
||||||
dump.rdb
|
dump.rdb
|
||||||
|
.vscode
|
||||||
435
app.py
435
app.py
@@ -1,8 +1,6 @@
|
|||||||
# dnsrecon-reduced/app.py
|
|
||||||
|
|
||||||
"""
|
"""
|
||||||
Flask application entry point for DNSRecon web interface.
|
Flask application entry point for DNSRecon web interface.
|
||||||
Provides REST API endpoints and serves the web interface with user session support.
|
Enhanced with user session management and task-based completion model.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import json
|
import json
|
||||||
@@ -11,44 +9,81 @@ from flask import Flask, render_template, request, jsonify, send_file, session
|
|||||||
from datetime import datetime, timezone, timedelta
|
from datetime import datetime, timezone, timedelta
|
||||||
import io
|
import io
|
||||||
|
|
||||||
from core.session_manager import session_manager
|
from core.session_manager import session_manager, UserIdentifier
|
||||||
from config import config
|
from config import config
|
||||||
from core.graph_manager import NodeType
|
|
||||||
from utils.helpers import is_valid_target
|
|
||||||
|
|
||||||
|
|
||||||
app = Flask(__name__)
|
app = Flask(__name__)
|
||||||
# Use centralized configuration for Flask settings
|
app.config['SECRET_KEY'] = 'dnsrecon-dev-key-change-in-production'
|
||||||
app.config['SECRET_KEY'] = config.flask_secret_key
|
app.config['PERMANENT_SESSION_LIFETIME'] = timedelta(hours=2) # 2 hour session lifetime
|
||||||
app.config['PERMANENT_SESSION_LIFETIME'] = timedelta(hours=config.flask_permanent_session_lifetime_hours)
|
|
||||||
|
|
||||||
def get_user_scanner():
|
def get_user_scanner():
|
||||||
"""
|
"""
|
||||||
Retrieves the scanner for the current session, or creates a new
|
Enhanced user scanner retrieval with user identification and session consolidation.
|
||||||
session and scanner if one doesn't exist.
|
Implements single session per user with seamless consolidation.
|
||||||
"""
|
"""
|
||||||
# Get current Flask session info for debugging
|
print("=== ENHANCED GET_USER_SCANNER ===")
|
||||||
current_flask_session_id = session.get('dnsrecon_session_id')
|
|
||||||
|
|
||||||
# Try to get existing session
|
try:
|
||||||
if current_flask_session_id:
|
# Extract user identification from request
|
||||||
existing_scanner = session_manager.get_session(current_flask_session_id)
|
client_ip, user_agent = UserIdentifier.extract_request_info(request)
|
||||||
if existing_scanner:
|
user_fingerprint = UserIdentifier.generate_user_fingerprint(client_ip, user_agent)
|
||||||
return current_flask_session_id, existing_scanner
|
|
||||||
|
print(f"User fingerprint: {user_fingerprint}")
|
||||||
# Create new session if none exists
|
print(f"Client IP: {client_ip}")
|
||||||
print("Creating new session as none was found...")
|
print(f"User Agent: {user_agent[:50]}...")
|
||||||
new_session_id = session_manager.create_session()
|
|
||||||
new_scanner = session_manager.get_session(new_session_id)
|
# Get current Flask session info for debugging
|
||||||
|
current_flask_session_id = session.get('dnsrecon_session_id')
|
||||||
if not new_scanner:
|
print(f"Flask session ID: {current_flask_session_id}")
|
||||||
raise Exception("Failed to create new scanner session")
|
|
||||||
|
# Try to get existing session first
|
||||||
# Store in Flask session
|
if current_flask_session_id:
|
||||||
session['dnsrecon_session_id'] = new_session_id
|
existing_scanner = session_manager.get_session(current_flask_session_id)
|
||||||
session.permanent = True
|
if existing_scanner:
|
||||||
|
# Verify session belongs to current user
|
||||||
return new_session_id, new_scanner
|
session_info = session_manager.get_session_info(current_flask_session_id)
|
||||||
|
if session_info.get('user_fingerprint') == user_fingerprint:
|
||||||
|
print(f"Found valid existing session {current_flask_session_id} for user {user_fingerprint}")
|
||||||
|
existing_scanner.session_id = current_flask_session_id
|
||||||
|
return current_flask_session_id, existing_scanner
|
||||||
|
else:
|
||||||
|
print(f"Session {current_flask_session_id} belongs to different user, will create new session")
|
||||||
|
else:
|
||||||
|
print(f"Session {current_flask_session_id} not found in Redis, will create new session")
|
||||||
|
|
||||||
|
# Create or replace user session (this handles consolidation automatically)
|
||||||
|
new_session_id = session_manager.create_or_replace_user_session(client_ip, user_agent)
|
||||||
|
new_scanner = session_manager.get_session(new_session_id)
|
||||||
|
|
||||||
|
if not new_scanner:
|
||||||
|
print(f"ERROR: Failed to retrieve newly created session {new_session_id}")
|
||||||
|
raise Exception("Failed to create new scanner session")
|
||||||
|
|
||||||
|
# Store in Flask session for browser persistence
|
||||||
|
session['dnsrecon_session_id'] = new_session_id
|
||||||
|
session.permanent = True
|
||||||
|
|
||||||
|
# Ensure session ID is set on scanner
|
||||||
|
new_scanner.session_id = new_session_id
|
||||||
|
|
||||||
|
# Get session info for user feedback
|
||||||
|
session_info = session_manager.get_session_info(new_session_id)
|
||||||
|
|
||||||
|
print(f"Session created/consolidated successfully")
|
||||||
|
print(f" - Session ID: {new_session_id}")
|
||||||
|
print(f" - User: {user_fingerprint}")
|
||||||
|
print(f" - Scanner status: {new_scanner.status}")
|
||||||
|
print(f" - Session age: {session_info.get('session_age_minutes', 0)} minutes")
|
||||||
|
|
||||||
|
return new_session_id, new_scanner
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
print(f"ERROR: Exception in get_user_scanner: {e}")
|
||||||
|
traceback.print_exc()
|
||||||
|
raise
|
||||||
|
|
||||||
|
|
||||||
@app.route('/')
|
@app.route('/')
|
||||||
def index():
|
def index():
|
||||||
@@ -59,72 +94,111 @@ def index():
|
|||||||
@app.route('/api/scan/start', methods=['POST'])
|
@app.route('/api/scan/start', methods=['POST'])
|
||||||
def start_scan():
|
def start_scan():
|
||||||
"""
|
"""
|
||||||
Start a new reconnaissance scan. Creates a new isolated scanner if
|
Start a new reconnaissance scan with enhanced user session management.
|
||||||
clear_graph is true, otherwise adds to the existing one.
|
|
||||||
"""
|
"""
|
||||||
print("=== API: /api/scan/start called ===")
|
print("=== API: /api/scan/start called ===")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
print("Getting JSON data from request...")
|
||||||
data = request.get_json()
|
data = request.get_json()
|
||||||
if not data or 'target' not in data:
|
print(f"Request data: {data}")
|
||||||
return jsonify({'success': False, 'error': 'Missing target in request'}), 400
|
|
||||||
|
|
||||||
target = data['target'].strip()
|
if not data or 'target_domain' not in data:
|
||||||
|
print("ERROR: Missing target_domain in request")
|
||||||
|
return jsonify({
|
||||||
|
'success': False,
|
||||||
|
'error': 'Missing target_domain in request'
|
||||||
|
}), 400
|
||||||
|
|
||||||
|
target_domain = data['target_domain'].strip()
|
||||||
max_depth = data.get('max_depth', config.default_recursion_depth)
|
max_depth = data.get('max_depth', config.default_recursion_depth)
|
||||||
clear_graph = data.get('clear_graph', True)
|
clear_graph = data.get('clear_graph', True)
|
||||||
force_rescan_target = data.get('force_rescan_target', None) # **FIX**: Get the new parameter
|
|
||||||
|
|
||||||
print(f"Parsed - target: '{target}', max_depth: {max_depth}, clear_graph: {clear_graph}, force_rescan: {force_rescan_target}")
|
print(f"Parsed - target_domain: '{target_domain}', max_depth: {max_depth}, clear_graph: {clear_graph}")
|
||||||
|
|
||||||
# Validation
|
# Validation
|
||||||
if not target:
|
if not target_domain:
|
||||||
return jsonify({'success': False, 'error': 'Target cannot be empty'}), 400
|
print("ERROR: Target domain cannot be empty")
|
||||||
if not is_valid_target(target):
|
return jsonify({
|
||||||
return jsonify({'success': False, 'error': 'Invalid target format. Please enter a valid domain or IP address.'}), 400
|
'success': False,
|
||||||
if not isinstance(max_depth, int) or not 1 <= max_depth <= 5:
|
'error': 'Target domain cannot be empty'
|
||||||
return jsonify({'success': False, 'error': 'Max depth must be an integer between 1 and 5'}), 400
|
}), 400
|
||||||
|
|
||||||
user_session_id, scanner = None, None
|
if not isinstance(max_depth, int) or max_depth < 1 or max_depth > 5:
|
||||||
|
print(f"ERROR: Invalid max_depth: {max_depth}")
|
||||||
if clear_graph:
|
return jsonify({
|
||||||
print("Clear graph requested: Creating a new, isolated scanner session.")
|
'success': False,
|
||||||
old_session_id = session.get('dnsrecon_session_id')
|
'error': 'Max depth must be an integer between 1 and 5'
|
||||||
if old_session_id:
|
}), 400
|
||||||
session_manager.terminate_session(old_session_id)
|
|
||||||
|
|
||||||
user_session_id = session_manager.create_session()
|
|
||||||
session['dnsrecon_session_id'] = user_session_id
|
|
||||||
session.permanent = True
|
|
||||||
scanner = session_manager.get_session(user_session_id)
|
|
||||||
else:
|
|
||||||
print("Adding to existing graph: Reusing the current scanner session.")
|
|
||||||
user_session_id, scanner = get_user_scanner()
|
|
||||||
|
|
||||||
if not scanner:
|
|
||||||
return jsonify({'success': False, 'error': 'Failed to get or create a scanner instance.'}), 500
|
|
||||||
|
|
||||||
print(f"Using scanner {id(scanner)} in session {user_session_id}")
|
print("Validation passed, getting user scanner...")
|
||||||
|
|
||||||
success = scanner.start_scan(target, max_depth, clear_graph=clear_graph, force_rescan_target=force_rescan_target) # **FIX**: Pass the new parameter
|
# Get user-specific scanner with enhanced session management
|
||||||
|
user_session_id, scanner = get_user_scanner()
|
||||||
|
|
||||||
|
# Ensure session ID is properly set
|
||||||
|
if not scanner.session_id:
|
||||||
|
scanner.session_id = user_session_id
|
||||||
|
|
||||||
|
print(f"Using session: {user_session_id}")
|
||||||
|
print(f"Scanner object ID: {id(scanner)}")
|
||||||
|
|
||||||
|
# Start scan
|
||||||
|
print(f"Calling start_scan on scanner {id(scanner)}...")
|
||||||
|
success = scanner.start_scan(target_domain, max_depth, clear_graph=clear_graph)
|
||||||
|
|
||||||
|
# Immediately update session state regardless of success
|
||||||
|
session_manager.update_session_scanner(user_session_id, scanner)
|
||||||
|
|
||||||
if success:
|
if success:
|
||||||
|
scan_session_id = scanner.logger.session_id
|
||||||
|
print(f"Scan started successfully with scan session ID: {scan_session_id}")
|
||||||
|
|
||||||
|
# Get session info for user feedback
|
||||||
|
session_info = session_manager.get_session_info(user_session_id)
|
||||||
|
|
||||||
return jsonify({
|
return jsonify({
|
||||||
'success': True,
|
'success': True,
|
||||||
'message': 'Scan started successfully',
|
'message': 'Scan started successfully',
|
||||||
'scan_id': scanner.logger.session_id,
|
'scan_id': scan_session_id,
|
||||||
'user_session_id': user_session_id,
|
'user_session_id': user_session_id,
|
||||||
|
'scanner_status': scanner.status,
|
||||||
|
'session_info': {
|
||||||
|
'user_fingerprint': session_info.get('user_fingerprint', 'unknown'),
|
||||||
|
'session_age_minutes': session_info.get('session_age_minutes', 0),
|
||||||
|
'consolidated': session_info.get('session_age_minutes', 0) > 0
|
||||||
|
},
|
||||||
|
'debug_info': {
|
||||||
|
'scanner_object_id': id(scanner),
|
||||||
|
'scanner_status': scanner.status
|
||||||
|
}
|
||||||
})
|
})
|
||||||
else:
|
else:
|
||||||
|
print("ERROR: Scanner returned False")
|
||||||
|
|
||||||
|
# Provide more detailed error information
|
||||||
|
error_details = {
|
||||||
|
'scanner_status': scanner.status,
|
||||||
|
'scanner_object_id': id(scanner),
|
||||||
|
'session_id': user_session_id,
|
||||||
|
'providers_count': len(scanner.providers) if hasattr(scanner, 'providers') else 0
|
||||||
|
}
|
||||||
|
|
||||||
return jsonify({
|
return jsonify({
|
||||||
'success': False,
|
'success': False,
|
||||||
'error': f'Failed to start scan (scanner status: {scanner.status})',
|
'error': f'Failed to start scan (scanner status: {scanner.status})',
|
||||||
|
'debug_info': error_details
|
||||||
}), 409
|
}), 409
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f"ERROR: Exception in start_scan endpoint: {e}")
|
print(f"ERROR: Exception in start_scan endpoint: {e}")
|
||||||
traceback.print_exc()
|
traceback.print_exc()
|
||||||
return jsonify({'success': False, 'error': f'Internal server error: {str(e)}'}), 500
|
return jsonify({
|
||||||
|
'success': False,
|
||||||
|
'error': f'Internal server error: {str(e)}'
|
||||||
|
}), 500
|
||||||
|
|
||||||
|
|
||||||
@app.route('/api/scan/stop', methods=['POST'])
|
@app.route('/api/scan/stop', methods=['POST'])
|
||||||
def stop_scan():
|
def stop_scan():
|
||||||
"""Stop the current scan with immediate GUI feedback."""
|
"""Stop the current scan with immediate GUI feedback."""
|
||||||
@@ -178,7 +252,7 @@ def stop_scan():
|
|||||||
|
|
||||||
@app.route('/api/scan/status', methods=['GET'])
|
@app.route('/api/scan/status', methods=['GET'])
|
||||||
def get_scan_status():
|
def get_scan_status():
|
||||||
"""Get current scan status with error handling."""
|
"""Get current scan status with enhanced session information."""
|
||||||
try:
|
try:
|
||||||
# Get user-specific scanner
|
# Get user-specific scanner
|
||||||
user_session_id, scanner = get_user_scanner()
|
user_session_id, scanner = get_user_scanner()
|
||||||
@@ -209,6 +283,15 @@ def get_scan_status():
|
|||||||
status = scanner.get_scan_status()
|
status = scanner.get_scan_status()
|
||||||
status['user_session_id'] = user_session_id
|
status['user_session_id'] = user_session_id
|
||||||
|
|
||||||
|
# Add enhanced session information
|
||||||
|
session_info = session_manager.get_session_info(user_session_id)
|
||||||
|
status['session_info'] = {
|
||||||
|
'user_fingerprint': session_info.get('user_fingerprint', 'unknown'),
|
||||||
|
'session_age_minutes': session_info.get('session_age_minutes', 0),
|
||||||
|
'client_ip': session_info.get('client_ip', 'unknown'),
|
||||||
|
'last_activity': session_info.get('last_activity')
|
||||||
|
}
|
||||||
|
|
||||||
# Additional debug info
|
# Additional debug info
|
||||||
status['debug_info'] = {
|
status['debug_info'] = {
|
||||||
'scanner_object_id': id(scanner),
|
'scanner_object_id': id(scanner),
|
||||||
@@ -237,7 +320,6 @@ def get_scan_status():
|
|||||||
}), 500
|
}), 500
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@app.route('/api/graph', methods=['GET'])
|
@app.route('/api/graph', methods=['GET'])
|
||||||
def get_graph_data():
|
def get_graph_data():
|
||||||
"""Get current graph data with error handling."""
|
"""Get current graph data with error handling."""
|
||||||
@@ -283,82 +365,6 @@ def get_graph_data():
|
|||||||
}), 500
|
}), 500
|
||||||
|
|
||||||
|
|
||||||
@app.route('/api/graph/node/<node_id>', methods=['DELETE'])
|
|
||||||
def delete_graph_node(node_id):
|
|
||||||
"""Delete a node from the graph for the current user session."""
|
|
||||||
try:
|
|
||||||
user_session_id, scanner = get_user_scanner()
|
|
||||||
if not scanner:
|
|
||||||
return jsonify({'success': False, 'error': 'No active session found'}), 404
|
|
||||||
|
|
||||||
success = scanner.graph.remove_node(node_id)
|
|
||||||
|
|
||||||
if success:
|
|
||||||
# Persist the change
|
|
||||||
session_manager.update_session_scanner(user_session_id, scanner)
|
|
||||||
return jsonify({'success': True, 'message': f'Node {node_id} deleted successfully.'})
|
|
||||||
else:
|
|
||||||
return jsonify({'success': False, 'error': f'Node {node_id} not found in graph.'}), 404
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
print(f"ERROR: Exception in delete_graph_node endpoint: {e}")
|
|
||||||
traceback.print_exc()
|
|
||||||
return jsonify({'success': False, 'error': f'Internal server error: {str(e)}'}), 500
|
|
||||||
|
|
||||||
|
|
||||||
@app.route('/api/graph/revert', methods=['POST'])
|
|
||||||
def revert_graph_action():
|
|
||||||
"""Reverts a graph action, such as re-adding a deleted node."""
|
|
||||||
try:
|
|
||||||
data = request.get_json()
|
|
||||||
if not data or 'type' not in data or 'data' not in data:
|
|
||||||
return jsonify({'success': False, 'error': 'Invalid revert request format'}), 400
|
|
||||||
|
|
||||||
user_session_id, scanner = get_user_scanner()
|
|
||||||
if not scanner:
|
|
||||||
return jsonify({'success': False, 'error': 'No active session found'}), 404
|
|
||||||
|
|
||||||
action_type = data['type']
|
|
||||||
action_data = data['data']
|
|
||||||
|
|
||||||
if action_type == 'delete':
|
|
||||||
# Re-add the node
|
|
||||||
node_to_add = action_data.get('node')
|
|
||||||
if node_to_add:
|
|
||||||
scanner.graph.add_node(
|
|
||||||
node_id=node_to_add['id'],
|
|
||||||
node_type=NodeType(node_to_add['type']),
|
|
||||||
attributes=node_to_add.get('attributes'),
|
|
||||||
description=node_to_add.get('description'),
|
|
||||||
metadata=node_to_add.get('metadata')
|
|
||||||
)
|
|
||||||
|
|
||||||
# Re-add the edges
|
|
||||||
edges_to_add = action_data.get('edges', [])
|
|
||||||
for edge in edges_to_add:
|
|
||||||
# Add edge only if both nodes exist to prevent errors
|
|
||||||
if scanner.graph.graph.has_node(edge['from']) and scanner.graph.graph.has_node(edge['to']):
|
|
||||||
scanner.graph.add_edge(
|
|
||||||
source_id=edge['from'],
|
|
||||||
target_id=edge['to'],
|
|
||||||
relationship_type=edge['metadata']['relationship_type'],
|
|
||||||
confidence_score=edge['metadata']['confidence_score'],
|
|
||||||
source_provider=edge['metadata']['source_provider'],
|
|
||||||
raw_data=edge.get('raw_data', {})
|
|
||||||
)
|
|
||||||
|
|
||||||
# Persist the change
|
|
||||||
session_manager.update_session_scanner(user_session_id, scanner)
|
|
||||||
return jsonify({'success': True, 'message': 'Delete action reverted successfully.'})
|
|
||||||
|
|
||||||
return jsonify({'success': False, 'error': f'Unknown revert action type: {action_type}'}), 400
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
print(f"ERROR: Exception in revert_graph_action endpoint: {e}")
|
|
||||||
traceback.print_exc()
|
|
||||||
return jsonify({'success': False, 'error': f'Internal server error: {str(e)}'}), 500
|
|
||||||
|
|
||||||
|
|
||||||
@app.route('/api/export', methods=['GET'])
|
@app.route('/api/export', methods=['GET'])
|
||||||
def export_results():
|
def export_results():
|
||||||
"""Export complete scan results as downloadable JSON for the user session."""
|
"""Export complete scan results as downloadable JSON for the user session."""
|
||||||
@@ -369,17 +375,22 @@ def export_results():
|
|||||||
# Get complete results
|
# Get complete results
|
||||||
results = scanner.export_results()
|
results = scanner.export_results()
|
||||||
|
|
||||||
# Add session information to export
|
# Add enhanced session information to export
|
||||||
|
session_info = session_manager.get_session_info(user_session_id)
|
||||||
results['export_metadata'] = {
|
results['export_metadata'] = {
|
||||||
'user_session_id': user_session_id,
|
'user_session_id': user_session_id,
|
||||||
|
'user_fingerprint': session_info.get('user_fingerprint', 'unknown'),
|
||||||
|
'client_ip': session_info.get('client_ip', 'unknown'),
|
||||||
|
'session_age_minutes': session_info.get('session_age_minutes', 0),
|
||||||
'export_timestamp': datetime.now(timezone.utc).isoformat(),
|
'export_timestamp': datetime.now(timezone.utc).isoformat(),
|
||||||
'export_type': 'user_session_results'
|
'export_type': 'user_session_results'
|
||||||
}
|
}
|
||||||
|
|
||||||
# Create filename with timestamp
|
# Create filename with user fingerprint
|
||||||
timestamp = datetime.now(timezone.utc).strftime('%Y%m%d_%H%M%S')
|
timestamp = datetime.now(timezone.utc).strftime('%Y%m%d_%H%M%S')
|
||||||
target = scanner.current_target or 'unknown'
|
target = scanner.current_target or 'unknown'
|
||||||
filename = f"dnsrecon_{target}_{timestamp}_{user_session_id[:8]}.json"
|
user_fp = session_info.get('user_fingerprint', 'unknown')[:8]
|
||||||
|
filename = f"dnsrecon_{target}_{timestamp}_{user_fp}.json"
|
||||||
|
|
||||||
# Create in-memory file
|
# Create in-memory file
|
||||||
json_data = json.dumps(results, indent=2, ensure_ascii=False)
|
json_data = json.dumps(results, indent=2, ensure_ascii=False)
|
||||||
@@ -404,18 +415,12 @@ def export_results():
|
|||||||
@app.route('/api/providers', methods=['GET'])
|
@app.route('/api/providers', methods=['GET'])
|
||||||
def get_providers():
|
def get_providers():
|
||||||
"""Get information about available providers for the user session."""
|
"""Get information about available providers for the user session."""
|
||||||
|
print("=== API: /api/providers called ===")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Get user-specific scanner
|
# Get user-specific scanner
|
||||||
user_session_id, scanner = get_user_scanner()
|
user_session_id, scanner = get_user_scanner()
|
||||||
|
|
||||||
if scanner:
|
|
||||||
completed_tasks = scanner.indicators_completed
|
|
||||||
enqueued_tasks = len(scanner.task_queue)
|
|
||||||
print(f"DEBUG: Tasks - Completed: {completed_tasks}, Enqueued: {enqueued_tasks}")
|
|
||||||
else:
|
|
||||||
print("DEBUG: No active scanner session found.")
|
|
||||||
|
|
||||||
provider_info = scanner.get_provider_info()
|
provider_info = scanner.get_provider_info()
|
||||||
|
|
||||||
return jsonify({
|
return jsonify({
|
||||||
@@ -490,6 +495,122 @@ def set_api_keys():
|
|||||||
'error': f'Internal server error: {str(e)}'
|
'error': f'Internal server error: {str(e)}'
|
||||||
}), 500
|
}), 500
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/api/session/info', methods=['GET'])
|
||||||
|
def get_session_info():
|
||||||
|
"""Get enhanced information about the current user session."""
|
||||||
|
try:
|
||||||
|
user_session_id, scanner = get_user_scanner()
|
||||||
|
session_info = session_manager.get_session_info(user_session_id)
|
||||||
|
|
||||||
|
return jsonify({
|
||||||
|
'success': True,
|
||||||
|
'session_info': session_info
|
||||||
|
})
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
print(f"ERROR: Exception in get_session_info endpoint: {e}")
|
||||||
|
traceback.print_exc()
|
||||||
|
return jsonify({
|
||||||
|
'success': False,
|
||||||
|
'error': f'Internal server error: {str(e)}'
|
||||||
|
}), 500
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/api/session/terminate', methods=['POST'])
|
||||||
|
def terminate_session():
|
||||||
|
"""Terminate the current user session."""
|
||||||
|
try:
|
||||||
|
user_session_id = session.get('dnsrecon_session_id')
|
||||||
|
|
||||||
|
if user_session_id:
|
||||||
|
success = session_manager.terminate_session(user_session_id)
|
||||||
|
# Clear Flask session
|
||||||
|
session.pop('dnsrecon_session_id', None)
|
||||||
|
|
||||||
|
return jsonify({
|
||||||
|
'success': success,
|
||||||
|
'message': 'Session terminated' if success else 'Session not found'
|
||||||
|
})
|
||||||
|
else:
|
||||||
|
return jsonify({
|
||||||
|
'success': False,
|
||||||
|
'error': 'No active session to terminate'
|
||||||
|
}), 400
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
print(f"ERROR: Exception in terminate_session endpoint: {e}")
|
||||||
|
traceback.print_exc()
|
||||||
|
return jsonify({
|
||||||
|
'success': False,
|
||||||
|
'error': f'Internal server error: {str(e)}'
|
||||||
|
}), 500
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/api/admin/sessions', methods=['GET'])
|
||||||
|
def list_sessions():
|
||||||
|
"""Admin endpoint to list all active sessions with enhanced information."""
|
||||||
|
try:
|
||||||
|
sessions = session_manager.list_active_sessions()
|
||||||
|
stats = session_manager.get_statistics()
|
||||||
|
|
||||||
|
return jsonify({
|
||||||
|
'success': True,
|
||||||
|
'sessions': sessions,
|
||||||
|
'statistics': stats
|
||||||
|
})
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
print(f"ERROR: Exception in list_sessions endpoint: {e}")
|
||||||
|
traceback.print_exc()
|
||||||
|
return jsonify({
|
||||||
|
'success': False,
|
||||||
|
'error': f'Internal server error: {str(e)}'
|
||||||
|
}), 500
|
||||||
|
|
||||||
|
|
||||||
|
@app.route('/api/health', methods=['GET'])
|
||||||
|
def health_check():
|
||||||
|
"""Health check endpoint with enhanced session statistics."""
|
||||||
|
try:
|
||||||
|
# Get session stats
|
||||||
|
session_stats = session_manager.get_statistics()
|
||||||
|
|
||||||
|
return jsonify({
|
||||||
|
'success': True,
|
||||||
|
'status': 'healthy',
|
||||||
|
'timestamp': datetime.now(timezone.utc).isoformat(),
|
||||||
|
'version': '2.0.0-enhanced',
|
||||||
|
'phase': 'enhanced_architecture',
|
||||||
|
'features': {
|
||||||
|
'multi_provider': True,
|
||||||
|
'concurrent_processing': True,
|
||||||
|
'real_time_updates': True,
|
||||||
|
'api_key_management': True,
|
||||||
|
'visualization': True,
|
||||||
|
'retry_logic': True,
|
||||||
|
'user_sessions': True,
|
||||||
|
'session_isolation': True,
|
||||||
|
'global_provider_caching': True,
|
||||||
|
'single_session_per_user': True,
|
||||||
|
'session_consolidation': True,
|
||||||
|
'task_completion_model': True
|
||||||
|
},
|
||||||
|
'session_statistics': session_stats,
|
||||||
|
'cache_info': {
|
||||||
|
'global_provider_cache': True,
|
||||||
|
'cache_location': '.cache/<provider_name>/',
|
||||||
|
'cache_expiry_hours': 12
|
||||||
|
}
|
||||||
|
})
|
||||||
|
except Exception as e:
|
||||||
|
print(f"ERROR: Exception in health_check endpoint: {e}")
|
||||||
|
return jsonify({
|
||||||
|
'success': False,
|
||||||
|
'error': f'Health check failed: {str(e)}'
|
||||||
|
}), 500
|
||||||
|
|
||||||
|
|
||||||
@app.errorhandler(404)
|
@app.errorhandler(404)
|
||||||
def not_found(error):
|
def not_found(error):
|
||||||
"""Handle 404 errors."""
|
"""Handle 404 errors."""
|
||||||
@@ -511,7 +632,7 @@ def internal_error(error):
|
|||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
print("Starting DNSRecon Flask application with user session support...")
|
print("Starting DNSRecon Flask application with enhanced user session support...")
|
||||||
|
|
||||||
# Load configuration from environment
|
# Load configuration from environment
|
||||||
config.load_from_env()
|
config.load_from_env()
|
||||||
|
|||||||
1
cache/crtsh/aibkurpark_de.json
vendored
1
cache/crtsh/aibkurpark_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/api_piped_mikoshi_de.json
vendored
1
cache/crtsh/api_piped_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"api.piped.mikoshi.de","first_cached":"2025-09-14T22:39:13.556737+00:00","last_upstream_query":"2025-09-15T15:40:40.243081+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":20189195635,"entry_timestamp":"2025-08-08T00:03:32.697","not_before":"2025-08-07T23:05:02","not_after":"2025-11-05T23:05:01","serial_number":"057a2344f2f0cff0a83bba587eb7326b4145","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":20189195721,"entry_timestamp":"2025-08-08T00:03:32.49","not_before":"2025-08-07T23:05:02","not_after":"2025-11-05T23:05:01","serial_number":"057a2344f2f0cff0a83bba587eb7326b4145","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":20060635277,"entry_timestamp":"2025-08-02T00:01:30.128","not_before":"2025-08-01T23:02:57","not_after":"2025-10-30T23:02:56","serial_number":"05a33355cd549118b90ec6c7219b2e283059","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":20060633958,"entry_timestamp":"2025-08-02T00:01:27.809","not_before":"2025-08-01T23:02:57","not_after":"2025-10-30T23:02:56","serial_number":"05a33355cd549118b90ec6c7219b2e283059","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":19929361217,"entry_timestamp":"2025-07-27T00:01:57.423","not_before":"2025-07-26T23:03:27","not_after":"2025-10-24T23:03:26","serial_number":"068f68538c5e6483213c6bdff93070960e6a","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":19929361084,"entry_timestamp":"2025-07-27T00:01:57.128","not_before":"2025-07-26T23:03:27","not_after":"2025-10-24T23:03:26","serial_number":"068f68538c5e6483213c6bdff93070960e6a","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":19803755935,"entry_timestamp":"2025-07-21T00:00:47.697","not_before":"2025-07-20T23:02:17","not_after":"2025-10-18T23:02:16","serial_number":"054cc0b2449a9ce1ce82324d844a1d4795f5","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":19803753732,"entry_timestamp":"2025-07-21T00:00:47.407","not_before":"2025-07-20T23:02:17","not_after":"2025-10-18T23:02:16","serial_number":"054cc0b2449a9ce1ce82324d844a1d4795f5","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":19668414027,"entry_timestamp":"2025-07-15T00:00:49.06","not_before":"2025-07-14T23:02:16","not_after":"2025-10-12T23:02:15","serial_number":"06ee7d4154b5a296e1cbefe4df2f6bfeba81","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":19668413763,"entry_timestamp":"2025-07-15T00:00:46.299","not_before":"2025-07-14T23:02:16","not_after":"2025-10-12T23:02:15","serial_number":"06ee7d4154b5a296e1cbefe4df2f6bfeba81","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":19523855233,"entry_timestamp":"2025-07-08T06:30:53.184","not_before":"2025-07-08T05:32:22","not_after":"2025-10-06T05:32:21","serial_number":"05956787611c0bc853c699866a6bc839eb89","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"api.piped.mikoshi.de","name_value":"api.piped.mikoshi.de","id":19523854846,"entry_timestamp":"2025-07-08T06:30:52.835","not_before":"2025-07-08T05:32:22","not_after":"2025-10-06T05:32:21","serial_number":"05956787611c0bc853c699866a6bc839eb89","result_count":2}]}
|
|
||||||
1
cache/crtsh/app_fleischkombinat-ost_de.json
vendored
1
cache/crtsh/app_fleischkombinat-ost_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"app.fleischkombinat-ost.de","first_cached":"2025-09-14T21:11:17.304989+00:00","last_upstream_query":"2025-09-15T09:15:09.273000+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"app.fleischkombinat-ost.de","name_value":"app.fleischkombinat-ost.de","id":19374493240,"entry_timestamp":"2025-07-01T14:09:36.354","not_before":"2025-07-01T13:11:00","not_after":"2025-09-29T13:10:59","serial_number":"0693231ff5e3212cabc2588e38b5d8337528","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"app.fleischkombinat-ost.de","name_value":"app.fleischkombinat-ost.de","id":19374489847,"entry_timestamp":"2025-07-01T14:09:30.117","not_before":"2025-07-01T13:11:00","not_after":"2025-09-29T13:10:59","serial_number":"0693231ff5e3212cabc2588e38b5d8337528","result_count":2}]}
|
|
||||||
1
cache/crtsh/bin_mikoshi_de.json
vendored
1
cache/crtsh/bin_mikoshi_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/c210lb_kasserver_com.json
vendored
1
cache/crtsh/c210lb_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"c210lb.kasserver.com","first_cached":"2025-09-15T11:49:15.879325+00:00","last_upstream_query":"2025-09-15T11:49:15.879328+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"test.c210lb.kasserver.com","name_value":"test.c210lb.kasserver.com\nwww.test.c210lb.kasserver.com","id":16080758209,"entry_timestamp":"2024-12-12T09:28:08.296","not_before":"2024-12-12T08:29:37","not_after":"2025-03-12T08:29:36","serial_number":"04d8425c5834ecd8914f58bb80e86bcae4a4","result_count":3},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"test.c210lb.kasserver.com","name_value":"test.c210lb.kasserver.com\nwww.test.c210lb.kasserver.com","id":15724227115,"entry_timestamp":"2024-12-12T09:28:07.974","not_before":"2024-12-12T08:29:37","not_after":"2025-03-12T08:29:36","serial_number":"04d8425c5834ecd8914f58bb80e86bcae4a4","result_count":3}]}
|
|
||||||
1
cache/crtsh/c214lb_kasserver_com.json
vendored
1
cache/crtsh/c214lb_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"c214lb.kasserver.com","first_cached":"2025-09-15T11:50:40.061642+00:00","last_upstream_query":"2025-09-15T11:50:40.061644+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"test.c214lb.kasserver.com","name_value":"test.c214lb.kasserver.com\nwww.test.c214lb.kasserver.com","id":16070500096,"entry_timestamp":"2024-12-11T14:25:06.88","not_before":"2024-12-11T13:26:36","not_after":"2025-03-11T13:26:35","serial_number":"047e10af0f4b68375cc37572f400322f1339","result_count":3},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"test.c214lb.kasserver.com","name_value":"test.c214lb.kasserver.com\nwww.test.c214lb.kasserver.com","id":15713437449,"entry_timestamp":"2024-12-11T14:25:06.342","not_before":"2024-12-11T13:26:36","not_after":"2025-03-11T13:26:35","serial_number":"047e10af0f4b68375cc37572f400322f1339","result_count":3}]}
|
|
||||||
1
cache/crtsh/cc24_dev.json
vendored
1
cache/crtsh/cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/cc24_mikoshi_de.json
vendored
1
cache/crtsh/cc24_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"cc24.mikoshi.de","first_cached":"2025-09-15T15:35:01.674879+00:00","last_upstream_query":"2025-09-15T15:35:01.674884+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":20275888640,"entry_timestamp":"2025-08-12T00:10:43.67","not_before":"2025-08-11T23:12:11","not_after":"2025-11-09T23:12:10","serial_number":"062bf40cddf77b5bb2efcc8d4394466e2f1b","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":20275882615,"entry_timestamp":"2025-08-12T00:10:41.383","not_before":"2025-08-11T23:12:11","not_after":"2025-11-09T23:12:10","serial_number":"062bf40cddf77b5bb2efcc8d4394466e2f1b","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":18987170126,"entry_timestamp":"2025-06-13T00:07:54.264","not_before":"2025-06-12T23:09:23","not_after":"2025-09-10T23:09:22","serial_number":"06146db771c214ec7474cb203b6ab75c21ae","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":18987169887,"entry_timestamp":"2025-06-13T00:07:53.982","not_before":"2025-06-12T23:09:23","not_after":"2025-09-10T23:09:22","serial_number":"06146db771c214ec7474cb203b6ab75c21ae","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":17831548102,"entry_timestamp":"2025-04-14T00:19:47.297","not_before":"2025-04-13T23:21:14","not_after":"2025-07-12T23:21:13","serial_number":"05d1504b6ed3a888b7577090d0f45a5e1f6e","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":17831584469,"entry_timestamp":"2025-04-14T00:19:45.047","not_before":"2025-04-13T23:21:14","not_after":"2025-07-12T23:21:13","serial_number":"05d1504b6ed3a888b7577090d0f45a5e1f6e","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":17170817151,"entry_timestamp":"2025-02-13T00:02:39.94","not_before":"2025-02-12T23:04:09","not_after":"2025-05-13T23:04:08","serial_number":"03b5b2d8ad135f1cf98224c917b797d0d844","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":16704121260,"entry_timestamp":"2025-02-13T00:02:39.313","not_before":"2025-02-12T23:04:09","not_after":"2025-05-13T23:04:08","serial_number":"03b5b2d8ad135f1cf98224c917b797d0d844","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":16672548148,"entry_timestamp":"2025-02-11T11:09:43.173","not_before":"2025-02-11T10:11:12","not_after":"2025-05-12T10:11:11","serial_number":"035ed55d94e5d12e255be2412e4035be70c9","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cc24.mikoshi.de","name_value":"cc24.mikoshi.de","id":16672548664,"entry_timestamp":"2025-02-11T11:09:42.207","not_before":"2025-02-11T10:11:12","not_after":"2025-05-12T10:11:11","serial_number":"035ed55d94e5d12e255be2412e4035be70c9","result_count":2}]}
|
|
||||||
1
cache/crtsh/cloud_aibkurpark_de.json
vendored
1
cache/crtsh/cloud_aibkurpark_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"cloud.aibkurpark.de","first_cached":"2025-09-14T22:46:40.289218+00:00","last_upstream_query":"2025-09-15T15:41:52.573914+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":20275873234,"entry_timestamp":"2025-08-12T00:10:04.8","not_before":"2025-08-11T23:11:34","not_after":"2025-11-09T23:11:33","serial_number":"0502f324ff45ca5c652de0f798011836a079","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":20275861222,"entry_timestamp":"2025-08-12T00:10:04.577","not_before":"2025-08-11T23:11:34","not_after":"2025-11-09T23:11:33","serial_number":"0502f324ff45ca5c652de0f798011836a079","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":18987158394,"entry_timestamp":"2025-06-13T00:07:19.317","not_before":"2025-06-12T23:08:44","not_after":"2025-09-10T23:08:43","serial_number":"051a486ebf950f3ccc46e6da6bc68a6477cc","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":18987157465,"entry_timestamp":"2025-06-13T00:07:15.025","not_before":"2025-06-12T23:08:44","not_after":"2025-09-10T23:08:43","serial_number":"051a486ebf950f3ccc46e6da6bc68a6477cc","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":17831539356,"entry_timestamp":"2025-04-14T00:19:09.511","not_before":"2025-04-13T23:20:39","not_after":"2025-07-12T23:20:38","serial_number":"05d6f819ef2fc26e17ae6bdc186dc1f83610","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":17831579073,"entry_timestamp":"2025-04-14T00:19:09.182","not_before":"2025-04-13T23:20:39","not_after":"2025-07-12T23:20:38","serial_number":"05d6f819ef2fc26e17ae6bdc186dc1f83610","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":17170816053,"entry_timestamp":"2025-02-13T00:02:33.08","not_before":"2025-02-12T23:04:01","not_after":"2025-05-13T23:04:00","serial_number":"0457fefa87af5335359d49883ebe4b6c2f22","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":16704122042,"entry_timestamp":"2025-02-13T00:02:31.914","not_before":"2025-02-12T23:04:01","not_after":"2025-05-13T23:04:00","serial_number":"0457fefa87af5335359d49883ebe4b6c2f22","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":17122421412,"entry_timestamp":"2025-02-09T15:23:02.809","not_before":"2025-02-09T14:24:32","not_after":"2025-05-10T14:24:31","serial_number":"0338963c04daad7c5c311c00a3b6c8999e36","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":16635575756,"entry_timestamp":"2025-02-09T15:23:02.704","not_before":"2025-02-09T14:24:32","not_after":"2025-05-10T14:24:31","serial_number":"0338963c04daad7c5c311c00a3b6c8999e36","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":11633733632,"entry_timestamp":"2024-01-03T19:39:23.66","not_before":"2024-01-03T18:39:23","not_after":"2024-04-02T18:39:22","serial_number":"040656c602c0e9460d045b95c5972319d195","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":11610236633,"entry_timestamp":"2024-01-03T19:39:23.48","not_before":"2024-01-03T18:39:23","not_after":"2024-04-02T18:39:22","serial_number":"040656c602c0e9460d045b95c5972319d195","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":10998276570,"entry_timestamp":"2023-11-04T20:08:52.9","not_before":"2023-11-04T19:08:52","not_after":"2024-02-02T19:08:51","serial_number":"03d6d33fb298a3b24e240893175dc0606f11","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"cloud.aibkurpark.de","name_value":"cloud.aibkurpark.de","id":10998275332,"entry_timestamp":"2023-11-04T20:08:52.78","not_before":"2023-11-04T19:08:52","not_after":"2024-02-02T19:08:51","serial_number":"03d6d33fb298a3b24e240893175dc0606f11","result_count":2}]}
|
|
||||||
1
cache/crtsh/cloud_cc24_dev.json
vendored
1
cache/crtsh/cloud_cc24_dev.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"cloud.cc24.dev","first_cached":"2025-09-14T21:36:46.109884+00:00","last_upstream_query":"2025-09-14T21:36:46.109891+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":20275852221,"entry_timestamp":"2025-08-12T00:08:47.671","not_before":"2025-08-11T23:10:16","not_after":"2025-11-09T23:10:15","serial_number":"0531b80da7039a455eb889201f8e62ba8cc9","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":20275837420,"entry_timestamp":"2025-08-12T00:08:47.014","not_before":"2025-08-11T23:10:16","not_after":"2025-11-09T23:10:15","serial_number":"0531b80da7039a455eb889201f8e62ba8cc9","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":18987132752,"entry_timestamp":"2025-06-13T00:06:00.445","not_before":"2025-06-12T23:07:30","not_after":"2025-09-10T23:07:29","serial_number":"066841f1e247045c3bb244d599955addbc66","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":18987133145,"entry_timestamp":"2025-06-13T00:06:00.158","not_before":"2025-06-12T23:07:30","not_after":"2025-09-10T23:07:29","serial_number":"066841f1e247045c3bb244d599955addbc66","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":17831462233,"entry_timestamp":"2025-04-14T00:17:56.109","not_before":"2025-04-13T23:19:24","not_after":"2025-07-12T23:19:23","serial_number":"051efd08e5e21db1fe47698ba7cb273c05b7","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":17831396685,"entry_timestamp":"2025-04-14T00:17:54.708","not_before":"2025-04-13T23:19:24","not_after":"2025-07-12T23:19:23","serial_number":"051efd08e5e21db1fe47698ba7cb273c05b7","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":17170810540,"entry_timestamp":"2025-02-13T00:02:07.841","not_before":"2025-02-12T23:03:37","not_after":"2025-05-13T23:03:36","serial_number":"03fe022033cd38b75215385397375a1a3741","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":16704105434,"entry_timestamp":"2025-02-13T00:02:07.708","not_before":"2025-02-12T23:03:37","not_after":"2025-05-13T23:03:36","serial_number":"03fe022033cd38b75215385397375a1a3741","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":17122390503,"entry_timestamp":"2025-02-09T15:21:20.386","not_before":"2025-02-09T14:22:49","not_after":"2025-05-10T14:22:48","serial_number":"031fcf4e1368ddbc276836806c2d3df6c376","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"cloud.cc24.dev","name_value":"cloud.cc24.dev","id":16635540435,"entry_timestamp":"2025-02-09T15:21:19.831","not_before":"2025-02-09T14:22:49","not_after":"2025-05-10T14:22:48","serial_number":"031fcf4e1368ddbc276836806c2d3df6c376","result_count":2}]}
|
|
||||||
1
cache/crtsh/cloud_mikoshi_de.json
vendored
1
cache/crtsh/cloud_mikoshi_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/code_cc24_dev.json
vendored
1
cache/crtsh/code_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/console_s3_cc24_dev.json
vendored
1
cache/crtsh/console_s3_cc24_dev.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"console.s3.cc24.dev","first_cached":"2025-09-14T21:33:25.149502+00:00","last_upstream_query":"2025-09-14T21:33:25.149505+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"console.s3.cc24.dev","name_value":"console.s3.cc24.dev","id":20287575466,"entry_timestamp":"2025-08-12T12:55:37.077","not_before":"2025-08-12T11:57:05","not_after":"2025-11-10T11:57:04","serial_number":"066bdfa83088f8d7e67284da94dd5d122ed6","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"console.s3.cc24.dev","name_value":"console.s3.cc24.dev","id":20287575457,"entry_timestamp":"2025-08-12T12:55:36.75","not_before":"2025-08-12T11:57:05","not_after":"2025-11-10T11:57:04","serial_number":"066bdfa83088f8d7e67284da94dd5d122ed6","result_count":2}]}
|
|
||||||
1
cache/crtsh/coturn_mikoshi_de.json
vendored
1
cache/crtsh/coturn_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"coturn.mikoshi.de","first_cached":"2025-09-14T22:37:33.747891+00:00","last_upstream_query":"2025-09-15T15:40:16.119867+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"coturn.mikoshi.de","name_value":"coturn.mikoshi.de","id":19208664152,"entry_timestamp":"2025-06-23T18:46:24.952","not_before":"2025-06-23T17:47:50","not_after":"2025-09-21T17:47:49","serial_number":"06d2bfd2d049388cf81d75548de9029700a8","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"coturn.mikoshi.de","name_value":"coturn.mikoshi.de","id":19208662980,"entry_timestamp":"2025-06-23T18:46:20.96","not_before":"2025-06-23T17:47:50","not_after":"2025-09-21T17:47:49","serial_number":"06d2bfd2d049388cf81d75548de9029700a8","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"coturn.mikoshi.de","name_value":"coturn.mikoshi.de","id":17345025382,"entry_timestamp":"2025-03-23T14:41:31.643","not_before":"2025-03-23T13:43:01","not_after":"2025-06-21T13:43:00","serial_number":"06245dcee4d40047ccbe0b4eac028e25edbe","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"coturn.mikoshi.de","name_value":"coturn.mikoshi.de","id":17345311273,"entry_timestamp":"2025-03-23T14:41:31.261","not_before":"2025-03-23T13:43:01","not_after":"2025-06-21T13:43:00","serial_number":"06245dcee4d40047ccbe0b4eac028e25edbe","result_count":2}]}
|
|
||||||
1
cache/crtsh/coturn_ms-it-services_de.json
vendored
1
cache/crtsh/coturn_ms-it-services_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"coturn.ms-it-services.de","first_cached":"2025-09-14T21:03:44.169328+00:00","last_upstream_query":"2025-09-15T19:08:39.390556+00:00","upstream_query_count":3,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"coturn.ms-it-services.de","name_value":"coturn.ms-it-services.de","id":14781803935,"entry_timestamp":"2024-10-03T09:53:12.473","not_before":"2024-10-03T08:54:42","not_after":"2025-01-01T08:54:41","serial_number":"0395c04e522a2715eebcb7fc4ffb3da1fdba","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"coturn.ms-it-services.de","name_value":"coturn.ms-it-services.de","id":14781794097,"entry_timestamp":"2024-10-03T09:53:12.142","not_before":"2024-10-03T08:54:42","not_after":"2025-01-01T08:54:41","serial_number":"0395c04e522a2715eebcb7fc4ffb3da1fdba","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"coturn.ms-it-services.de","name_value":"coturn.ms-it-services.de","id":14773518990,"entry_timestamp":"2024-10-02T19:20:49.687","not_before":"2024-10-02T18:22:19","not_after":"2024-12-31T18:22:18","serial_number":"04f26242ac1b2ac659ac2e19ae2522ce3274","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"coturn.ms-it-services.de","name_value":"coturn.ms-it-services.de","id":14773501988,"entry_timestamp":"2024-10-02T19:20:49.356","not_before":"2024-10-02T18:22:19","not_after":"2024-12-31T18:22:18","serial_number":"04f26242ac1b2ac659ac2e19ae2522ce3274","result_count":2}]}
|
|
||||||
1
cache/crtsh/dd10032_kasserver_com.json
vendored
1
cache/crtsh/dd10032_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd10032.kasserver.com","first_cached":"2025-09-15T11:49:39.835993+00:00","last_upstream_query":"2025-09-15T11:49:39.835996+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"wernertest.dd10032.kasserver.com","name_value":"wernertest.dd10032.kasserver.com\nwww.wernertest.dd10032.kasserver.com","id":19900983152,"entry_timestamp":"2025-07-25T15:08:27.328","not_before":"2025-07-25T14:09:55","not_after":"2025-10-23T14:09:54","serial_number":"053bbb3a5a38ad78a0298a2d7de184471db7","result_count":3},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"wernertest.dd10032.kasserver.com","name_value":"wernertest.dd10032.kasserver.com\nwww.wernertest.dd10032.kasserver.com","id":19900983133,"entry_timestamp":"2025-07-25T15:08:25.843","not_before":"2025-07-25T14:09:55","not_after":"2025-10-23T14:09:54","serial_number":"053bbb3a5a38ad78a0298a2d7de184471db7","result_count":3},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"wernertest.dd10032.kasserver.com","name_value":"wernertest.dd10032.kasserver.com\nwww.wernertest.dd10032.kasserver.com","id":18625008970,"entry_timestamp":"2025-05-26T02:06:09.384","not_before":"2025-05-26T01:07:39","not_after":"2025-08-24T01:07:38","serial_number":"05b1920c57feb37d80ab28e0a745fd4e5711","result_count":3},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"wernertest.dd10032.kasserver.com","name_value":"wernertest.dd10032.kasserver.com\nwww.wernertest.dd10032.kasserver.com","id":18625008687,"entry_timestamp":"2025-05-26T02:06:09.149","not_before":"2025-05-26T01:07:39","not_after":"2025-08-24T01:07:38","serial_number":"05b1920c57feb37d80ab28e0a745fd4e5711","result_count":3},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"wernertest.dd10032.kasserver.com","name_value":"wernertest.dd10032.kasserver.com\nwww.wernertest.dd10032.kasserver.com","id":17676635727,"entry_timestamp":"2025-03-18T15:24:41.197","not_before":"2025-03-18T14:26:10","not_after":"2025-06-16T14:26:09","serial_number":"05f4a3810f24e3916e04d2b092c864e92b04","result_count":3},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"wernertest.dd10032.kasserver.com","name_value":"wernertest.dd10032.kasserver.com\nwww.wernertest.dd10032.kasserver.com","id":17214728959,"entry_timestamp":"2025-03-18T15:24:40.996","not_before":"2025-03-18T14:26:10","not_after":"2025-06-16T14:26:09","serial_number":"05f4a3810f24e3916e04d2b092c864e92b04","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd15000_kasserver_com.json
vendored
1
cache/crtsh/dd15000_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd15000.kasserver.com","first_cached":"2025-09-15T11:49:51.832307+00:00","last_upstream_query":"2025-09-15T11:49:51.832313+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"litera.dd15000.kasserver.com","name_value":"litera.dd15000.kasserver.com\nwww.litera.dd15000.kasserver.com","id":17806830,"entry_timestamp":"2016-05-03T09:27:49.724","not_before":"2016-05-03T08:28:00","not_after":"2016-08-01T08:28:00","serial_number":"030cb7992941a2b2860858446a989170f2db","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd17120_kasserver_com.json
vendored
1
cache/crtsh/dd17120_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd17120.kasserver.com","first_cached":"2025-09-15T11:50:52.134823+00:00","last_upstream_query":"2025-09-15T11:50:52.134826+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"letest.dd17120.kasserver.com","name_value":"letest.dd17120.kasserver.com\nwww.letest.dd17120.kasserver.com","id":3820560694,"entry_timestamp":"2020-12-23T15:35:32.403","not_before":"2020-12-23T14:35:32","not_after":"2021-03-23T14:35:32","serial_number":"04c92961dbb2d095ce5690f41ab75431abed","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"letest.dd17120.kasserver.com","name_value":"letest.dd17120.kasserver.com\nwww.letest.dd17120.kasserver.com","id":3820561611,"entry_timestamp":"2020-12-23T15:35:32.275","not_before":"2020-12-23T14:35:32","not_after":"2021-03-23T14:35:32","serial_number":"04c92961dbb2d095ce5690f41ab75431abed","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"letest.dd17120.kasserver.com","name_value":"letest.dd17120.kasserver.com\nwww.letest.dd17120.kasserver.com","id":3820541026,"entry_timestamp":"2020-12-23T15:29:20.032","not_before":"2020-12-23T14:29:19","not_after":"2021-03-23T14:29:19","serial_number":"03446bb8f1a87b4f0cb40fd1030125240eed","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"letest.dd17120.kasserver.com","name_value":"letest.dd17120.kasserver.com\nwww.letest.dd17120.kasserver.com","id":3820541140,"entry_timestamp":"2020-12-23T15:29:19.847","not_before":"2020-12-23T14:29:19","not_after":"2021-03-23T14:29:19","serial_number":"03446bb8f1a87b4f0cb40fd1030125240eed","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"letest.dd17120.kasserver.com","name_value":"letest.dd17120.kasserver.com\nwww.letest.dd17120.kasserver.com","id":3820538580,"entry_timestamp":"2020-12-23T15:28:46.532","not_before":"2020-12-23T14:28:46","not_after":"2021-03-23T14:28:46","serial_number":"03c3eb90d764b2c2a0ac04beed2e876f03e1","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"letest.dd17120.kasserver.com","name_value":"letest.dd17120.kasserver.com\nwww.letest.dd17120.kasserver.com","id":3820538531,"entry_timestamp":"2020-12-23T15:28:46.312","not_before":"2020-12-23T14:28:46","not_after":"2021-03-23T14:28:46","serial_number":"03c3eb90d764b2c2a0ac04beed2e876f03e1","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"letest.dd17120.kasserver.com","name_value":"letest.dd17120.kasserver.com\nwww.letest.dd17120.kasserver.com","id":3820537101,"entry_timestamp":"2020-12-23T15:28:12.742","not_before":"2020-12-23T14:28:12","not_after":"2021-03-23T14:28:12","serial_number":"034cd7432ae3dba0e76509de60666d58f1e4","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"letest.dd17120.kasserver.com","name_value":"letest.dd17120.kasserver.com\nwww.letest.dd17120.kasserver.com","id":3820537306,"entry_timestamp":"2020-12-23T15:28:12.459","not_before":"2020-12-23T14:28:12","not_after":"2021-03-23T14:28:12","serial_number":"034cd7432ae3dba0e76509de60666d58f1e4","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd18904_kasserver_com.json
vendored
1
cache/crtsh/dd18904_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd18904.kasserver.com","first_cached":"2025-09-15T11:49:03.689534+00:00","last_upstream_query":"2025-09-15T11:49:03.689540+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"thomas.dd18904.kasserver.com","name_value":"thomas.dd18904.kasserver.com\nwww.thomas.dd18904.kasserver.com","id":6766495494,"entry_timestamp":"2022-05-20T07:44:34.589","not_before":"2022-05-20T06:44:34","not_after":"2022-08-18T06:44:33","serial_number":"046fc2606fb3d5e6438be8586ff20e296e8e","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"thomas.dd18904.kasserver.com","name_value":"thomas.dd18904.kasserver.com\nwww.thomas.dd18904.kasserver.com","id":6766489032,"entry_timestamp":"2022-05-20T07:44:34.335","not_before":"2022-05-20T06:44:34","not_after":"2022-08-18T06:44:33","serial_number":"046fc2606fb3d5e6438be8586ff20e296e8e","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"thomas.dd18904.kasserver.com","name_value":"thomas.dd18904.kasserver.com\nwww.thomas.dd18904.kasserver.com","id":6766485621,"entry_timestamp":"2022-05-20T07:42:24.52","not_before":"2022-05-20T06:42:24","not_after":"2022-08-18T06:42:23","serial_number":"03a52deebc041d7b6b4246ec8fc0455345a8","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"thomas.dd18904.kasserver.com","name_value":"thomas.dd18904.kasserver.com\nwww.thomas.dd18904.kasserver.com","id":6766485609,"entry_timestamp":"2022-05-20T07:42:24.413","not_before":"2022-05-20T06:42:24","not_after":"2022-08-18T06:42:23","serial_number":"03a52deebc041d7b6b4246ec8fc0455345a8","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd19536_kasserver_com.json
vendored
1
cache/crtsh/dd19536_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd19536.kasserver.com","first_cached":"2025-09-15T11:47:57.902219+00:00","last_upstream_query":"2025-09-15T11:47:57.902224+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"login.jitsi.dd19536.kasserver.com","name_value":"login.jitsi.dd19536.kasserver.com\nwww.login.jitsi.dd19536.kasserver.com","id":14154546767,"entry_timestamp":"2024-08-12T05:52:06.917","not_before":"2024-08-12T04:53:36","not_after":"2024-11-10T04:53:35","serial_number":"04fca7ec58a44a09702875c5d2c6c9d6401b","result_count":3},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"login.jitsi.dd19536.kasserver.com","name_value":"login.jitsi.dd19536.kasserver.com\nwww.login.jitsi.dd19536.kasserver.com","id":14108590090,"entry_timestamp":"2024-08-12T05:52:06.531","not_before":"2024-08-12T04:53:36","not_after":"2024-11-10T04:53:35","serial_number":"04fca7ec58a44a09702875c5d2c6c9d6401b","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd24432_kasserver_com.json
vendored
1
cache/crtsh/dd24432_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd24432.kasserver.com","first_cached":"2025-09-15T11:51:28.290098+00:00","last_upstream_query":"2025-09-15T11:51:28.290102+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"test1.dd24432.kasserver.com","name_value":"test1.dd24432.kasserver.com\nwww.test1.dd24432.kasserver.com","id":157863863,"entry_timestamp":"2017-06-20T11:17:00.314","not_before":"2017-06-20T10:17:00","not_after":"2017-09-18T10:17:00","serial_number":"03c833715204594e6946b183652f0be2055e","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd27030_kasserver_com.json
vendored
1
cache/crtsh/dd27030_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd27030.kasserver.com","first_cached":"2025-09-15T11:48:39.556778+00:00","last_upstream_query":"2025-09-15T11:48:39.556783+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"wernertest.dd27030.kasserver.com","name_value":"wernertest.dd27030.kasserver.com\nwww.wernertest.dd27030.kasserver.com","id":17324952949,"entry_timestamp":"2025-03-19T12:30:16.583","not_before":"2025-03-19T11:31:44","not_after":"2025-06-17T11:31:43","serial_number":"0686091c483d6c1a55f241d41c29523fc3f0","result_count":3},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"wernertest.dd27030.kasserver.com","name_value":"wernertest.dd27030.kasserver.com\nwww.wernertest.dd27030.kasserver.com","id":17314228622,"entry_timestamp":"2025-03-19T12:30:14.317","not_before":"2025-03-19T11:31:44","not_after":"2025-06-17T11:31:43","serial_number":"0686091c483d6c1a55f241d41c29523fc3f0","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd27120_kasserver_com.json
vendored
1
cache/crtsh/dd27120_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd27120.kasserver.com","first_cached":"2025-09-15T11:47:51.800693+00:00","last_upstream_query":"2025-09-15T11:47:51.800697+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"otrs2.dd27120.kasserver.com","name_value":"otrs2.dd27120.kasserver.com\nwww.otrs2.dd27120.kasserver.com","id":1476617768,"entry_timestamp":"2019-05-09T10:40:20.848","not_before":"2019-05-09T09:40:19","not_after":"2019-08-07T09:40:19","serial_number":"0342616d062ef6933a703e4de279da4e38b6","result_count":3},{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"otrs2.dd27120.kasserver.com","name_value":"otrs2.dd27120.kasserver.com\nwww.otrs2.dd27120.kasserver.com","id":1458597695,"entry_timestamp":"2019-05-09T10:40:20.006","not_before":"2019-05-09T09:40:19","not_after":"2019-08-07T09:40:19","serial_number":"0342616d062ef6933a703e4de279da4e38b6","result_count":3},{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"otrs2.dd27120.kasserver.com","name_value":"otrs2.dd27120.kasserver.com\nwww.otrs2.dd27120.kasserver.com","id":1476589088,"entry_timestamp":"2019-05-09T10:17:27.514","not_before":"2019-05-09T09:17:26","not_after":"2019-08-07T09:17:26","serial_number":"0398319827847d1df54260d7273cd5ea6d5c","result_count":3},{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"otrs2.dd27120.kasserver.com","name_value":"otrs2.dd27120.kasserver.com\nwww.otrs2.dd27120.kasserver.com","id":1457806572,"entry_timestamp":"2019-05-09T10:17:26.68","not_before":"2019-05-09T09:17:26","not_after":"2019-08-07T09:17:26","serial_number":"0398319827847d1df54260d7273cd5ea6d5c","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd27932_kasserver_com.json
vendored
1
cache/crtsh/dd27932_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd27932.kasserver.com","first_cached":"2025-09-15T11:48:51.718464+00:00","last_upstream_query":"2025-09-15T11:48:51.718468+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"switest.dd27932.kasserver.com","name_value":"switest.dd27932.kasserver.com\nwww.switest.dd27932.kasserver.com","id":99317583,"entry_timestamp":"2017-03-05T01:00:32.21","not_before":"2017-03-05T00:01:00","not_after":"2017-06-03T00:01:00","serial_number":"03e0f1fb8ff2a8a742ee45c2da39e4bd68ed","result_count":3},{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"switest.dd27932.kasserver.com","name_value":"switest.dd27932.kasserver.com\nwww.switest.dd27932.kasserver.com","id":69451932,"entry_timestamp":"2016-12-21T10:01:06.051","not_before":"2016-12-21T09:01:00","not_after":"2017-03-21T09:01:00","serial_number":"03a4f79d5ed3c6661f6f2c430a1a1546e0d8","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd28826_kasserver_com.json
vendored
1
cache/crtsh/dd28826_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd28826.kasserver.com","first_cached":"2025-09-15T11:50:28.011163+00:00","last_upstream_query":"2025-09-15T11:50:28.011167+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"161gs86116.dd28826.kasserver.com","name_value":"161gs86116.dd28826.kasserver.com\nwww.161gs86116.dd28826.kasserver.com","id":138535930,"entry_timestamp":"2017-05-15T11:15:13.509","not_before":"2017-05-15T10:15:00","not_after":"2017-08-13T10:15:00","serial_number":"03bca24f2de29889f514c04523a1f8945136","result_count":3},{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"161gs86116.dd28826.kasserver.com","name_value":"161gs86116.dd28826.kasserver.com\nwww.161gs86116.dd28826.kasserver.com","id":40903630,"entry_timestamp":"2016-10-07T12:00:36.009","not_before":"2016-10-07T11:01:00","not_after":"2017-01-05T11:01:00","serial_number":"0385d74f0a39572de61c2b7765c94d848c82","result_count":3},{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"161gs86116.dd28826.kasserver.com","name_value":"161gs86116.dd28826.kasserver.com\nwww.161gs86116.dd28826.kasserver.com","id":40901166,"entry_timestamp":"2016-10-07T11:54:34.702","not_before":"2016-10-07T10:55:00","not_after":"2017-01-05T10:55:00","serial_number":"039ef12598cb7aaf33c04d4c90983b0829fe","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd30002_kasserver_com.json
vendored
1
cache/crtsh/dd30002_kasserver_com.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/dd39314_kasserver_com.json
vendored
1
cache/crtsh/dd39314_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd39314.kasserver.com","first_cached":"2025-09-15T11:47:59.899335+00:00","last_upstream_query":"2025-09-15T11:47:59.899339+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"http2.dd39314.kasserver.com","name_value":"http2.dd39314.kasserver.com\nwww.http2.dd39314.kasserver.com","id":35834556,"entry_timestamp":"2016-09-27T11:27:08.538","not_before":"2016-09-27T10:27:00","not_after":"2016-12-26T10:27:00","serial_number":"03024be246b9452e3e5e7dbed7e502a1ee16","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd39910_kasserver_com.json
vendored
1
cache/crtsh/dd39910_kasserver_com.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/dd44624_kasserver_com.json
vendored
1
cache/crtsh/dd44624_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd44624.kasserver.com","first_cached":"2025-09-15T11:50:16.030946+00:00","last_upstream_query":"2025-09-15T11:50:16.030952+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"aitest.dd44624.kasserver.com","name_value":"aitest.dd44624.kasserver.com\nwww.aitest.dd44624.kasserver.com","id":2366400329,"entry_timestamp":"2020-01-21T09:28:49.996","not_before":"2020-01-21T08:28:49","not_after":"2020-04-20T08:28:49","serial_number":"0319ee129ecf09b2a08a98d6cd7d92062efe","result_count":3},{"issuer_ca_id":16418,"issuer_name":"C=US, O=Let's Encrypt, CN=Let's Encrypt Authority X3","common_name":"aitest.dd44624.kasserver.com","name_value":"aitest.dd44624.kasserver.com\nwww.aitest.dd44624.kasserver.com","id":2355073809,"entry_timestamp":"2020-01-21T09:28:49.389","not_before":"2020-01-21T08:28:49","not_after":"2020-04-20T08:28:49","serial_number":"0319ee129ecf09b2a08a98d6cd7d92062efe","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd46900_kasserver_com.json
vendored
1
cache/crtsh/dd46900_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd46900.kasserver.com","first_cached":"2025-09-15T11:49:27.811675+00:00","last_upstream_query":"2025-09-15T11:49:27.811679+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"wernertest.dd46900.kasserver.com","name_value":"wernertest.dd46900.kasserver.com\nwww.wernertest.dd46900.kasserver.com","id":19891708696,"entry_timestamp":"2025-07-25T04:48:43.91","not_before":"2025-07-25T03:50:12","not_after":"2025-10-23T03:50:11","serial_number":"06ab9c27b7c779ee95cb82b66f985d87601c","result_count":3},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"wernertest.dd46900.kasserver.com","name_value":"wernertest.dd46900.kasserver.com\nwww.wernertest.dd46900.kasserver.com","id":19891706706,"entry_timestamp":"2025-07-25T04:48:42.387","not_before":"2025-07-25T03:50:12","not_after":"2025-10-23T03:50:11","serial_number":"06ab9c27b7c779ee95cb82b66f985d87601c","result_count":3},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"wernertest.dd46900.kasserver.com","name_value":"wernertest.dd46900.kasserver.com\nwww.wernertest.dd46900.kasserver.com","id":18618700336,"entry_timestamp":"2025-05-25T17:22:29.664","not_before":"2025-05-25T16:23:59","not_after":"2025-08-23T16:23:58","serial_number":"05abb39369b76694a079e1109a93e616bf18","result_count":3},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"wernertest.dd46900.kasserver.com","name_value":"wernertest.dd46900.kasserver.com\nwww.wernertest.dd46900.kasserver.com","id":18618704556,"entry_timestamp":"2025-05-25T17:22:29.449","not_before":"2025-05-25T16:23:59","not_after":"2025-08-23T16:23:58","serial_number":"05abb39369b76694a079e1109a93e616bf18","result_count":3},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"wernertest.dd46900.kasserver.com","name_value":"wernertest.dd46900.kasserver.com\nwww.wernertest.dd46900.kasserver.com","id":17676537327,"entry_timestamp":"2025-03-18T15:18:10.621","not_before":"2025-03-18T14:19:39","not_after":"2025-06-16T14:19:38","serial_number":"0610cb215ef16b2619af24f76349e349939c","result_count":3},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"wernertest.dd46900.kasserver.com","name_value":"wernertest.dd46900.kasserver.com\nwww.wernertest.dd46900.kasserver.com","id":17230420365,"entry_timestamp":"2025-03-18T15:18:09.941","not_before":"2025-03-18T14:19:39","not_after":"2025-06-16T14:19:38","serial_number":"0610cb215ef16b2619af24f76349e349939c","result_count":3}]}
|
|
||||||
1
cache/crtsh/dd55600_kasserver_com.json
vendored
1
cache/crtsh/dd55600_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dd55600.kasserver.com","first_cached":"2025-09-15T11:47:55.878914+00:00","last_upstream_query":"2025-09-15T11:47:55.878918+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"letest.dd55600.kasserver.com","name_value":"letest.dd55600.kasserver.com\nwww.letest.dd55600.kasserver.com","id":20031222371,"entry_timestamp":"2025-07-31T13:43:52.442","not_before":"2025-07-31T12:45:20","not_after":"2025-10-29T12:45:19","serial_number":"06eb0a6229e52eabee01d67ad7a999665794","result_count":3},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"letest.dd55600.kasserver.com","name_value":"letest.dd55600.kasserver.com\nwww.letest.dd55600.kasserver.com","id":20031228125,"entry_timestamp":"2025-07-31T13:43:50.126","not_before":"2025-07-31T12:45:20","not_after":"2025-10-29T12:45:19","serial_number":"06eb0a6229e52eabee01d67ad7a999665794","result_count":3}]}
|
|
||||||
1
cache/crtsh/deck0_mikoshi_de.json
vendored
1
cache/crtsh/deck0_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"deck0.mikoshi.de","first_cached":"2025-09-15T15:38:47.110406+00:00","last_upstream_query":"2025-09-15T15:38:47.110410+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":20275849021,"entry_timestamp":"2025-08-12T00:08:09.613","not_before":"2025-08-11T23:09:39","not_after":"2025-11-09T23:09:38","serial_number":"05785ec3742887857667078bfa75b21eda78","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":20275847401,"entry_timestamp":"2025-08-12T00:08:09.412","not_before":"2025-08-11T23:09:39","not_after":"2025-11-09T23:09:38","serial_number":"05785ec3742887857667078bfa75b21eda78","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":18987122782,"entry_timestamp":"2025-06-13T00:05:25.356","not_before":"2025-06-12T23:06:50","not_after":"2025-09-10T23:06:49","serial_number":"051c77c6054cfd6720e600ff7f5a9a42b40f","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":18987120711,"entry_timestamp":"2025-06-13T00:05:21.025","not_before":"2025-06-12T23:06:50","not_after":"2025-09-10T23:06:49","serial_number":"051c77c6054cfd6720e600ff7f5a9a42b40f","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":17831463793,"entry_timestamp":"2025-04-14T00:17:17.936","not_before":"2025-04-13T23:18:47","not_after":"2025-07-12T23:18:46","serial_number":"0652fad83eb95b52eb2a47457872b0408fef","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":17831393060,"entry_timestamp":"2025-04-14T00:17:17.692","not_before":"2025-04-13T23:18:47","not_after":"2025-07-12T23:18:46","serial_number":"0652fad83eb95b52eb2a47457872b0408fef","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":16704809501,"entry_timestamp":"2025-02-13T00:02:01.475","not_before":"2025-02-12T23:03:29","not_after":"2025-05-13T23:03:28","serial_number":"03336511cc977e31ed4534ee782a8b856e03","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":16704111156,"entry_timestamp":"2025-02-13T00:01:59.346","not_before":"2025-02-12T23:03:29","not_after":"2025-05-13T23:03:28","serial_number":"03336511cc977e31ed4534ee782a8b856e03","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":17122385653,"entry_timestamp":"2025-02-09T15:20:45.345","not_before":"2025-02-09T14:22:14","not_after":"2025-05-10T14:22:13","serial_number":"04b3ee603f5f877a0acbb32ff5e68570fc4e","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck0.mikoshi.de","name_value":"deck0.mikoshi.de","id":16636455622,"entry_timestamp":"2025-02-09T15:20:45.058","not_before":"2025-02-09T14:22:14","not_after":"2025-05-10T14:22:13","serial_number":"04b3ee603f5f877a0acbb32ff5e68570fc4e","result_count":2}]}
|
|
||||||
1
cache/crtsh/deck_mikoshi_de.json
vendored
1
cache/crtsh/deck_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"deck.mikoshi.de","first_cached":"2025-09-15T15:37:15.209223+00:00","last_upstream_query":"2025-09-15T15:37:15.209226+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":20296619407,"entry_timestamp":"2025-08-13T00:04:28.809","not_before":"2025-08-12T23:05:56","not_after":"2025-11-10T23:05:55","serial_number":"068288286ab2af41b51f2a6f0e4c6d6343ed","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":20296618002,"entry_timestamp":"2025-08-13T00:04:26.633","not_before":"2025-08-12T23:05:56","not_after":"2025-11-10T23:05:55","serial_number":"068288286ab2af41b51f2a6f0e4c6d6343ed","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":19007501639,"entry_timestamp":"2025-06-14T00:02:02.957","not_before":"2025-06-13T23:03:32","not_after":"2025-09-11T23:03:31","serial_number":"065d212df70f76cc2d2e5fb1cd3372785d45","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":19007529630,"entry_timestamp":"2025-06-14T00:02:02.71","not_before":"2025-06-13T23:03:32","not_after":"2025-09-11T23:03:31","serial_number":"065d212df70f76cc2d2e5fb1cd3372785d45","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":17831396750,"entry_timestamp":"2025-04-14T00:16:41.797","not_before":"2025-04-13T23:18:10","not_after":"2025-07-12T23:18:09","serial_number":"063db5541056907cd7d1a6dd6b90337970cf","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":17831377445,"entry_timestamp":"2025-04-14T00:16:41.02","not_before":"2025-04-13T23:18:10","not_after":"2025-07-12T23:18:09","serial_number":"063db5541056907cd7d1a6dd6b90337970cf","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":16704809458,"entry_timestamp":"2025-02-13T00:01:53.117","not_before":"2025-02-12T23:03:22","not_after":"2025-05-13T23:03:21","serial_number":"04dd85a6d9cafec66b0b7ca3753191d347fa","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":16704110323,"entry_timestamp":"2025-02-13T00:01:52.905","not_before":"2025-02-12T23:03:22","not_after":"2025-05-13T23:03:21","serial_number":"04dd85a6d9cafec66b0b7ca3753191d347fa","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":16636696795,"entry_timestamp":"2025-02-09T15:20:11.318","not_before":"2025-02-09T14:21:40","not_after":"2025-05-10T14:21:39","serial_number":"043f12b3e0be22dd52133924bbb5c1ded954","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"deck.mikoshi.de","name_value":"deck.mikoshi.de","id":16635534286,"entry_timestamp":"2025-02-09T15:20:10.545","not_before":"2025-02-09T14:21:40","not_after":"2025-05-10T14:21:39","serial_number":"043f12b3e0be22dd52133924bbb5c1ded954","result_count":2}]}
|
|
||||||
1
cache/crtsh/dev_mikoshi_de.json
vendored
1
cache/crtsh/dev_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dev.mikoshi.de","first_cached":"2025-09-15T15:39:15.810886+00:00","last_upstream_query":"2025-09-15T15:39:15.810888+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":20275786291,"entry_timestamp":"2025-08-12T00:05:23.212","not_before":"2025-08-11T23:06:52","not_after":"2025-11-09T23:06:51","serial_number":"062b3368dce240f65370e3fcf5b868f3cb16","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":20275788995,"entry_timestamp":"2025-08-12T00:05:22.899","not_before":"2025-08-11T23:06:52","not_after":"2025-11-09T23:06:51","serial_number":"062b3368dce240f65370e3fcf5b868f3cb16","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":18987037224,"entry_timestamp":"2025-06-13T00:03:00.049","not_before":"2025-06-12T23:04:25","not_after":"2025-09-10T23:04:24","serial_number":"05dac28fc29cc4ffa07ebb5d652b5d3c220d","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":18987083431,"entry_timestamp":"2025-06-13T00:02:56.025","not_before":"2025-06-12T23:04:25","not_after":"2025-09-10T23:04:24","serial_number":"05dac28fc29cc4ffa07ebb5d652b5d3c220d","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":17831380009,"entry_timestamp":"2025-04-14T00:13:37.609","not_before":"2025-04-13T23:15:06","not_after":"2025-07-12T23:15:05","serial_number":"05acf9ab2ed9a117ef386d05d6cfae3ed258","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":17831368386,"entry_timestamp":"2025-04-14T00:13:36.704","not_before":"2025-04-13T23:15:06","not_after":"2025-07-12T23:15:05","serial_number":"05acf9ab2ed9a117ef386d05d6cfae3ed258","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":17170799512,"entry_timestamp":"2025-02-13T00:01:03.867","not_before":"2025-02-12T23:02:33","not_after":"2025-05-13T23:02:32","serial_number":"04324a24cab88153022dd4a82cd03a1d5bec","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":16705238915,"entry_timestamp":"2025-02-13T00:01:03.58","not_before":"2025-02-12T23:02:33","not_after":"2025-05-13T23:02:32","serial_number":"04324a24cab88153022dd4a82cd03a1d5bec","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":16635493913,"entry_timestamp":"2025-02-09T15:16:09.908","not_before":"2025-02-09T14:17:39","not_after":"2025-05-10T14:17:38","serial_number":"03567895f69539097a4621b1b2f68c04f75c","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"dev.mikoshi.de","name_value":"dev.mikoshi.de","id":16635494768,"entry_timestamp":"2025-02-09T15:16:09.457","not_before":"2025-02-09T14:17:39","not_after":"2025-05-10T14:17:38","serial_number":"03567895f69539097a4621b1b2f68c04f75c","result_count":2}]}
|
|
||||||
1
cache/crtsh/dnsrecon_cc24_dev.json
vendored
1
cache/crtsh/dnsrecon_cc24_dev.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"dnsrecon.cc24.dev","first_cached":"2025-09-14T21:33:58.773156+00:00","last_upstream_query":"2025-09-14T21:33:58.773159+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295809,"issuer_name":"C=US, O=Let's Encrypt, CN=E8","common_name":"dnsrecon.cc24.dev","name_value":"dnsrecon.cc24.dev","id":20965278266,"entry_timestamp":"2025-09-12T09:49:23.247","not_before":"2025-09-12T08:50:53","not_after":"2025-12-11T08:50:52","serial_number":"060fbe619a364febd85aebccb1c6fcf7153f","result_count":2},{"issuer_ca_id":295809,"issuer_name":"C=US, O=Let's Encrypt, CN=E8","common_name":"dnsrecon.cc24.dev","name_value":"dnsrecon.cc24.dev","id":20965277886,"entry_timestamp":"2025-09-12T09:49:23.039","not_before":"2025-09-12T08:50:53","not_after":"2025-12-11T08:50:52","serial_number":"060fbe619a364febd85aebccb1c6fcf7153f","result_count":2}]}
|
|
||||||
1
cache/crtsh/einkauf_mikoshi_de.json
vendored
1
cache/crtsh/einkauf_mikoshi_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/electrum_mikoshi_de.json
vendored
1
cache/crtsh/electrum_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"electrum.mikoshi.de","first_cached":"2025-09-15T15:35:50.844792+00:00","last_upstream_query":"2025-09-15T15:35:50.844796+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":20275738086,"entry_timestamp":"2025-08-12T00:01:00.455","not_before":"2025-08-11T23:02:29","not_after":"2025-11-09T23:02:28","serial_number":"0607acd92d2e4ed2c21a0fd8879b4fcc1da6","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":20275720683,"entry_timestamp":"2025-08-12T00:01:00.121","not_before":"2025-08-11T23:02:29","not_after":"2025-11-09T23:02:28","serial_number":"0607acd92d2e4ed2c21a0fd8879b4fcc1da6","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":18987049091,"entry_timestamp":"2025-06-13T00:00:34.9","not_before":"2025-06-12T23:02:04","not_after":"2025-09-10T23:02:03","serial_number":"05b5b53d270451a0c1287c9aad800bf21a5b","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":18987049255,"entry_timestamp":"2025-06-13T00:00:34.45","not_before":"2025-06-12T23:02:04","not_after":"2025-09-10T23:02:03","serial_number":"05b5b53d270451a0c1287c9aad800bf21a5b","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":17831223577,"entry_timestamp":"2025-04-14T00:03:09.379","not_before":"2025-04-13T23:04:37","not_after":"2025-07-12T23:04:36","serial_number":"06f83ec1efdf5ec3387c8f241234448b1618","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":17831306766,"entry_timestamp":"2025-04-14T00:03:07.489","not_before":"2025-04-13T23:04:37","not_after":"2025-07-12T23:04:36","serial_number":"06f83ec1efdf5ec3387c8f241234448b1618","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":17170790341,"entry_timestamp":"2025-02-13T00:00:15.841","not_before":"2025-02-12T23:01:45","not_after":"2025-05-13T23:01:44","serial_number":"0400d022ec8e2ded90606b9276eef71a1664","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":16704088726,"entry_timestamp":"2025-02-13T00:00:15.527","not_before":"2025-02-12T23:01:45","not_after":"2025-05-13T23:01:44","serial_number":"0400d022ec8e2ded90606b9276eef71a1664","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":16635426761,"entry_timestamp":"2025-02-09T15:12:05.797","not_before":"2025-02-09T14:13:35","not_after":"2025-05-10T14:13:34","serial_number":"045de320229498cb69676d307ec9467fa83e","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"electrum.mikoshi.de","name_value":"electrum.mikoshi.de","id":16635425561,"entry_timestamp":"2025-02-09T15:12:05.539","not_before":"2025-02-09T14:13:35","not_after":"2025-05-10T14:13:34","serial_number":"045de320229498cb69676d307ec9467fa83e","result_count":2}]}
|
|
||||||
1
cache/crtsh/element_cc24_dev.json
vendored
1
cache/crtsh/element_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/fleischkombinat-ost_de.json
vendored
1
cache/crtsh/fleischkombinat-ost_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"fleischkombinat-ost.de","first_cached":"2025-09-14T21:11:00.028593+00:00","last_upstream_query":"2025-09-15T09:15:04.796648+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"app.fleischkombinat-ost.de","name_value":"app.fleischkombinat-ost.de","id":19374493240,"entry_timestamp":"2025-07-01T14:09:36.354","not_before":"2025-07-01T13:11:00","not_after":"2025-09-29T13:10:59","serial_number":"0693231ff5e3212cabc2588e38b5d8337528","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"app.fleischkombinat-ost.de","name_value":"app.fleischkombinat-ost.de","id":19374489847,"entry_timestamp":"2025-07-01T14:09:30.117","not_before":"2025-07-01T13:11:00","not_after":"2025-09-29T13:10:59","serial_number":"0693231ff5e3212cabc2588e38b5d8337528","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"fleischkombinat-ost.de","name_value":"fleischkombinat-ost.de","id":19374378473,"entry_timestamp":"2025-07-01T14:01:50.593","not_before":"2025-07-01T13:03:20","not_after":"2025-09-29T13:03:19","serial_number":"06315dfed8c93d1497c26b21c448857b6f2c","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"fleischkombinat-ost.de","name_value":"fleischkombinat-ost.de","id":19374376791,"entry_timestamp":"2025-07-01T14:01:50.385","not_before":"2025-07-01T13:03:20","not_after":"2025-09-29T13:03:19","serial_number":"06315dfed8c93d1497c26b21c448857b6f2c","result_count":2},{"issuer_ca_id":158800,"issuer_name":"C=AT, O=ZeroSSL, CN=ZeroSSL RSA Domain Secure Site CA","common_name":"*.fleischkombinat-ost.de","name_value":"*.fleischkombinat-ost.de\nfleischkombinat-ost.de","id":19369530786,"entry_timestamp":"2025-07-01T09:12:02.496","not_before":"2025-07-01T00:00:00","not_after":"2025-09-29T23:59:59","serial_number":"07c51a2c164b3a6c5769b0e03a9f4085","result_count":3},{"issuer_ca_id":158800,"issuer_name":"C=AT, O=ZeroSSL, CN=ZeroSSL RSA Domain Secure Site CA","common_name":"*.fleischkombinat-ost.de","name_value":"*.fleischkombinat-ost.de\nfleischkombinat-ost.de","id":19369530780,"entry_timestamp":"2025-07-01T09:12:01.04","not_before":"2025-07-01T00:00:00","not_after":"2025-09-29T23:59:59","serial_number":"07c51a2c164b3a6c5769b0e03a9f4085","result_count":3}]}
|
|
||||||
1
cache/crtsh/forensics_cc24_dev.json
vendored
1
cache/crtsh/forensics_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/forum_cc24_dev.json
vendored
1
cache/crtsh/forum_cc24_dev.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"forum.cc24.dev","first_cached":"2025-09-14T21:37:55.208070+00:00","last_upstream_query":"2025-09-14T21:37:55.208073+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":20275855672,"entry_timestamp":"2025-08-12T00:09:25.872","not_before":"2025-08-11T23:10:53","not_after":"2025-11-09T23:10:52","serial_number":"05e7fa80df3e45a7ecec44d25dafad0904e5","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":20275850214,"entry_timestamp":"2025-08-12T00:09:23.776","not_before":"2025-08-11T23:10:53","not_after":"2025-11-09T23:10:52","serial_number":"05e7fa80df3e45a7ecec44d25dafad0904e5","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":18987146858,"entry_timestamp":"2025-06-13T00:06:40.302","not_before":"2025-06-12T23:08:03","not_after":"2025-09-10T23:08:02","serial_number":"055efe65125ee83f14545a2f5d99590e635f","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":18987145127,"entry_timestamp":"2025-06-13T00:06:34.097","not_before":"2025-06-12T23:08:03","not_after":"2025-09-10T23:08:02","serial_number":"055efe65125ee83f14545a2f5d99590e635f","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":17831472866,"entry_timestamp":"2025-04-14T00:18:33.004","not_before":"2025-04-13T23:20:02","not_after":"2025-07-12T23:20:01","serial_number":"064fefe879ff5e732ce6c6a63f1e911db568","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":17831402949,"entry_timestamp":"2025-04-14T00:18:32.224","not_before":"2025-04-13T23:20:02","not_after":"2025-07-12T23:20:01","serial_number":"064fefe879ff5e732ce6c6a63f1e911db568","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":17170813968,"entry_timestamp":"2025-02-13T00:02:16.332","not_before":"2025-02-12T23:03:44","not_after":"2025-05-13T23:03:43","serial_number":"036e8e11288b54228648cab399477ea43c56","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":16705244862,"entry_timestamp":"2025-02-13T00:02:14.119","not_before":"2025-02-12T23:03:44","not_after":"2025-05-13T23:03:43","serial_number":"036e8e11288b54228648cab399477ea43c56","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":17122401304,"entry_timestamp":"2025-02-09T15:21:53.869","not_before":"2025-02-09T14:23:23","not_after":"2025-05-10T14:23:22","serial_number":"03e0928358e7636b3cc59a7dbc3b188bd7ca","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"forum.cc24.dev","name_value":"forum.cc24.dev","id":16636457115,"entry_timestamp":"2025-02-09T15:21:53.738","not_before":"2025-02-09T14:23:23","not_after":"2025-05-10T14:23:22","serial_number":"03e0928358e7636b3cc59a7dbc3b188bd7ca","result_count":2}]}
|
|
||||||
1
cache/crtsh/git_cc24_dev.json
vendored
1
cache/crtsh/git_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/git_mikoshi_de.json
vendored
1
cache/crtsh/git_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"git.mikoshi.de","first_cached":"2025-09-14T22:41:53.620008+00:00","last_upstream_query":"2025-09-15T15:34:53.671880+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295817,"issuer_name":"C=US, O=Let's Encrypt, CN=R13","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":20885174925,"entry_timestamp":"2025-09-09T00:00:36.192","not_before":"2025-09-08T23:02:04","not_after":"2025-12-07T23:02:03","serial_number":"0615e6baefd21c126c2cdddf6dcb2507d56f","result_count":2},{"issuer_ca_id":295817,"issuer_name":"C=US, O=Let's Encrypt, CN=R13","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":20885174361,"entry_timestamp":"2025-09-09T00:00:34.086","not_before":"2025-09-08T23:02:04","not_after":"2025-12-07T23:02:03","serial_number":"0615e6baefd21c126c2cdddf6dcb2507d56f","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":19559742374,"entry_timestamp":"2025-07-10T00:00:34.37","not_before":"2025-07-09T23:02:04","not_after":"2025-10-07T23:02:03","serial_number":"06d9b4e5782769647aa336db245fff31f620","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":19559742305,"entry_timestamp":"2025-07-10T00:00:34.165","not_before":"2025-07-09T23:02:04","not_after":"2025-10-07T23:02:03","serial_number":"06d9b4e5782769647aa336db245fff31f620","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":18325953073,"entry_timestamp":"2025-05-10T12:59:59.883","not_before":"2025-05-10T12:01:29","not_after":"2025-08-08T12:01:28","serial_number":"06b8ea37731e6560b6c67edbefbacfd4105c","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":18325952875,"entry_timestamp":"2025-05-10T12:59:59.504","not_before":"2025-05-10T12:01:29","not_after":"2025-08-08T12:01:28","serial_number":"06b8ea37731e6560b6c67edbefbacfd4105c","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":17831317349,"entry_timestamp":"2025-04-14T00:13:00.421","not_before":"2025-04-13T23:14:28","not_after":"2025-07-12T23:14:27","serial_number":"056bf6f5e9efb81381e56fbd6e28c32f4f90","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":17831304428,"entry_timestamp":"2025-04-14T00:12:58.339","not_before":"2025-04-13T23:14:28","not_after":"2025-07-12T23:14:27","serial_number":"056bf6f5e9efb81381e56fbd6e28c32f4f90","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":17170797746,"entry_timestamp":"2025-02-13T00:00:56.931","not_before":"2025-02-12T23:02:26","not_after":"2025-05-13T23:02:25","serial_number":"03b7b4c4c6233e74e3ad23a7bb2e455fcc3f","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":16705238631,"entry_timestamp":"2025-02-13T00:00:56.689","not_before":"2025-02-12T23:02:26","not_after":"2025-05-13T23:02:25","serial_number":"03b7b4c4c6233e74e3ad23a7bb2e455fcc3f","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":17122277510,"entry_timestamp":"2025-02-09T15:15:36.076","not_before":"2025-02-09T14:17:05","not_after":"2025-05-10T14:17:04","serial_number":"0404f208620f990403046876d32991e8f320","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"git.mikoshi.de","name_value":"git.mikoshi.de","id":16635498764,"entry_timestamp":"2025-02-09T15:15:35.493","not_before":"2025-02-09T14:17:05","not_after":"2025-05-10T14:17:04","serial_number":"0404f208620f990403046876d32991e8f320","result_count":2}]}
|
|
||||||
1
cache/crtsh/graph_cc24_dev.json
vendored
1
cache/crtsh/graph_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/hoarder_cc24_dev.json
vendored
1
cache/crtsh/hoarder_cc24_dev.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"hoarder.cc24.dev","first_cached":"2025-09-14T21:33:21.961821+00:00","last_upstream_query":"2025-09-14T21:33:21.961824+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"hoarder.cc24.dev","name_value":"hoarder.cc24.dev","id":17651896110,"entry_timestamp":"2025-04-03T19:39:13.874","not_before":"2025-04-03T18:40:43","not_after":"2025-07-02T18:40:42","serial_number":"062e2d17ef9c31ca560ab40299e0e00701c0","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"hoarder.cc24.dev","name_value":"hoarder.cc24.dev","id":17610697718,"entry_timestamp":"2025-04-03T19:39:13.522","not_before":"2025-04-03T18:40:43","not_after":"2025-07-02T18:40:42","serial_number":"062e2d17ef9c31ca560ab40299e0e00701c0","result_count":2}]}
|
|
||||||
1
cache/crtsh/hub_cc24_dev.json
vendored
1
cache/crtsh/hub_cc24_dev.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"hub.cc24.dev","first_cached":"2025-09-14T21:33:48.364903+00:00","last_upstream_query":"2025-09-14T21:33:48.364908+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":20317572962,"entry_timestamp":"2025-08-14T00:02:25.791","not_before":"2025-08-13T23:03:55","not_after":"2025-11-11T23:03:54","serial_number":"05ee38e8ae7bff4c5c414784e2c64d933b7a","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":20317572564,"entry_timestamp":"2025-08-14T00:02:25.581","not_before":"2025-08-13T23:03:55","not_after":"2025-11-11T23:03:54","serial_number":"05ee38e8ae7bff4c5c414784e2c64d933b7a","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":20189230503,"entry_timestamp":"2025-08-08T00:05:51.4","not_before":"2025-08-07T23:07:21","not_after":"2025-11-05T23:07:20","serial_number":"0502ddbfff49d3df12f3a15ba33ad895ccf1","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":20189230585,"entry_timestamp":"2025-08-08T00:05:51.234","not_before":"2025-08-07T23:07:21","not_after":"2025-11-05T23:07:20","serial_number":"0502ddbfff49d3df12f3a15ba33ad895ccf1","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":20060643580,"entry_timestamp":"2025-08-02T00:02:15.895","not_before":"2025-08-01T23:03:45","not_after":"2025-10-30T23:03:44","serial_number":"05c750b95573daa27e1a55a8803a2e6d21ad","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":20060643807,"entry_timestamp":"2025-08-02T00:02:15.683","not_before":"2025-08-01T23:03:45","not_after":"2025-10-30T23:03:44","serial_number":"05c750b95573daa27e1a55a8803a2e6d21ad","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":19929376671,"entry_timestamp":"2025-07-27T00:02:55.144","not_before":"2025-07-26T23:04:22","not_after":"2025-10-24T23:04:21","serial_number":"052c4fe632cf76d0306310cfb9faecc224ba","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":19929375724,"entry_timestamp":"2025-07-27T00:02:52.759","not_before":"2025-07-26T23:04:22","not_after":"2025-10-24T23:04:21","serial_number":"052c4fe632cf76d0306310cfb9faecc224ba","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":19803769083,"entry_timestamp":"2025-07-21T00:01:15.171","not_before":"2025-07-20T23:02:40","not_after":"2025-10-18T23:02:39","serial_number":"05e9927e9ea80c233e8a9502f11e4958cd06","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":19803778513,"entry_timestamp":"2025-07-21T00:01:10.987","not_before":"2025-07-20T23:02:40","not_after":"2025-10-18T23:02:39","serial_number":"05e9927e9ea80c233e8a9502f11e4958cd06","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":19665369127,"entry_timestamp":"2025-07-14T21:15:11.092","not_before":"2025-07-14T20:16:38","not_after":"2025-10-12T20:16:37","serial_number":"069e06aa4855496ab6b766889d6b45135d4e","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"hub.cc24.dev","name_value":"hub.cc24.dev","id":19665373526,"entry_timestamp":"2025-07-14T21:15:08.866","not_before":"2025-07-14T20:16:38","not_after":"2025-10-12T20:16:37","serial_number":"069e06aa4855496ab6b766889d6b45135d4e","result_count":2}]}
|
|
||||||
1
cache/crtsh/ip_mikoshi_de.json
vendored
1
cache/crtsh/ip_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"ip.mikoshi.de","first_cached":"2025-09-15T15:35:38.687945+00:00","last_upstream_query":"2025-09-15T15:35:38.687948+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":20275799673,"entry_timestamp":"2025-08-12T00:06:09.895","not_before":"2025-08-11T23:07:39","not_after":"2025-11-09T23:07:38","serial_number":"05abe97c54f28c570ed093f2c2064d2bd909","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":20275798755,"entry_timestamp":"2025-08-12T00:06:09.726","not_before":"2025-08-11T23:07:39","not_after":"2025-11-09T23:07:38","serial_number":"05abe97c54f28c570ed093f2c2064d2bd909","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":18987090979,"entry_timestamp":"2025-06-13T00:03:33.968","not_before":"2025-06-12T23:05:03","not_after":"2025-09-10T23:05:02","serial_number":"05861ba42c1b941a781e56836a1286805486","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":18987090443,"entry_timestamp":"2025-06-13T00:03:33.547","not_before":"2025-06-12T23:05:03","not_after":"2025-09-10T23:05:02","serial_number":"05861ba42c1b941a781e56836a1286805486","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":17831378289,"entry_timestamp":"2025-04-14T00:14:14.055","not_before":"2025-04-13T23:15:43","not_after":"2025-07-12T23:15:42","serial_number":"05e8a5d6c84779035ff19c804b3eb98d997a","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":17831323161,"entry_timestamp":"2025-04-14T00:14:13.821","not_before":"2025-04-13T23:15:43","not_after":"2025-07-12T23:15:42","serial_number":"05e8a5d6c84779035ff19c804b3eb98d997a","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":17170801048,"entry_timestamp":"2025-02-13T00:01:10.43","not_before":"2025-02-12T23:02:40","not_after":"2025-05-13T23:02:39","serial_number":"03af85f30dcc4dbca816456437ccc8760356","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":16704107826,"entry_timestamp":"2025-02-13T00:01:10.264","not_before":"2025-02-12T23:02:40","not_after":"2025-05-13T23:02:39","serial_number":"03af85f30dcc4dbca816456437ccc8760356","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":17122289078,"entry_timestamp":"2025-02-09T15:16:43.147","not_before":"2025-02-09T14:18:12","not_after":"2025-05-10T14:18:11","serial_number":"03ef24688a55744de221b025bac45c4898c5","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"ip.mikoshi.de","name_value":"ip.mikoshi.de","id":16635485733,"entry_timestamp":"2025-02-09T15:16:43.019","not_before":"2025-02-09T14:18:12","not_after":"2025-05-10T14:18:11","serial_number":"03ef24688a55744de221b025bac45c4898c5","result_count":2}]}
|
|
||||||
1
cache/crtsh/kasserver_com.json
vendored
1
cache/crtsh/kasserver_com.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/keep_cc24_dev.json
vendored
1
cache/crtsh/keep_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/kvmtest2_kasserver_com.json
vendored
1
cache/crtsh/kvmtest2_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"kvmtest2.kasserver.com","first_cached":"2025-09-15T11:48:15.510962+00:00","last_upstream_query":"2025-09-15T11:48:15.510965+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmtest2.kasserver.com","name_value":"nextcloud.kvmtest2.kasserver.com\nwww.nextcloud.kvmtest2.kasserver.com","id":11872742578,"entry_timestamp":"2024-01-26T08:07:49.508","not_before":"2024-01-26T07:07:49","not_after":"2024-04-25T07:07:48","serial_number":"033a4ebc78836fabf49336d55237d61606b1","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmtest2.kasserver.com","name_value":"nextcloud.kvmtest2.kasserver.com\nwww.nextcloud.kvmtest2.kasserver.com","id":11872748114,"entry_timestamp":"2024-01-26T08:07:49.316","not_before":"2024-01-26T07:07:49","not_after":"2024-04-25T07:07:48","serial_number":"033a4ebc78836fabf49336d55237d61606b1","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmtest2.kasserver.com","name_value":"nextcloud.kvmtest2.kasserver.com\nwww.nextcloud.kvmtest2.kasserver.com","id":11867647719,"entry_timestamp":"2024-01-25T21:52:18.981","not_before":"2024-01-25T20:52:18","not_after":"2024-04-24T20:52:17","serial_number":"03e7aa261a2cd46d7de1f7a140b4a798eb21","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmtest2.kasserver.com","name_value":"nextcloud.kvmtest2.kasserver.com\nwww.nextcloud.kvmtest2.kasserver.com","id":11867647704,"entry_timestamp":"2024-01-25T21:52:18.671","not_before":"2024-01-25T20:52:18","not_after":"2024-04-24T20:52:17","serial_number":"03e7aa261a2cd46d7de1f7a140b4a798eb21","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmtest2.kasserver.com","name_value":"nextcloud.kvmtest2.kasserver.com\nwww.nextcloud.kvmtest2.kasserver.com","id":11864252076,"entry_timestamp":"2024-01-25T14:44:38.589","not_before":"2024-01-25T13:44:38","not_after":"2024-04-24T13:44:37","serial_number":"03542eac1a93d9a614ab26535b70ea26a6a5","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmtest2.kasserver.com","name_value":"nextcloud.kvmtest2.kasserver.com\nwww.nextcloud.kvmtest2.kasserver.com","id":11864248841,"entry_timestamp":"2024-01-25T14:44:38.505","not_before":"2024-01-25T13:44:38","not_after":"2024-04-24T13:44:37","serial_number":"03542eac1a93d9a614ab26535b70ea26a6a5","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmtest2.kasserver.com","name_value":"nextcloud.kvmtest2.kasserver.com\nwww.nextcloud.kvmtest2.kasserver.com","id":11853852174,"entry_timestamp":"2024-01-24T14:39:08.051","not_before":"2024-01-24T13:39:06","not_after":"2024-04-23T13:39:05","serial_number":"047335e6361f4c5477b088ff8cd2719d4d9a","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmtest2.kasserver.com","name_value":"nextcloud.kvmtest2.kasserver.com\nwww.nextcloud.kvmtest2.kasserver.com","id":11853852409,"entry_timestamp":"2024-01-24T14:39:07.46","not_before":"2024-01-24T13:39:06","not_after":"2024-04-23T13:39:05","serial_number":"047335e6361f4c5477b088ff8cd2719d4d9a","result_count":3}]}
|
|
||||||
1
cache/crtsh/kvmthkr_kasserver_com.json
vendored
1
cache/crtsh/kvmthkr_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"kvmthkr.kasserver.com","first_cached":"2025-09-15T11:50:03.952251+00:00","last_upstream_query":"2025-09-15T11:50:03.952256+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":105484,"issuer_name":"C=GB, ST=Greater Manchester, L=Salford, O=Sectigo Limited, CN=Sectigo ECC Domain Validation Secure Server CA","common_name":"ec.kvmthkr.kasserver.com","name_value":"ec.kvmthkr.kasserver.com\nwww.ec.kvmthkr.kasserver.com","id":14939726486,"entry_timestamp":"2024-10-15T07:02:39.33","not_before":"2024-10-15T00:00:00","not_after":"2024-11-14T23:59:59","serial_number":"6827c362ad3905799d6a865b0f1f60f8","result_count":3},{"issuer_ca_id":105484,"issuer_name":"C=GB, ST=Greater Manchester, L=Salford, O=Sectigo Limited, CN=Sectigo ECC Domain Validation Secure Server CA","common_name":"ec.kvmthkr.kasserver.com","name_value":"ec.kvmthkr.kasserver.com\nwww.ec.kvmthkr.kasserver.com","id":14939726540,"entry_timestamp":"2024-10-15T07:02:38.057","not_before":"2024-10-15T00:00:00","not_after":"2024-11-14T23:59:59","serial_number":"6827c362ad3905799d6a865b0f1f60f8","result_count":3}]}
|
|
||||||
1
cache/crtsh/kvmu12s1_kasserver_com.json
vendored
1
cache/crtsh/kvmu12s1_kasserver_com.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"kvmu12s1.kasserver.com","first_cached":"2025-09-15T11:51:04.220452+00:00","last_upstream_query":"2025-09-15T11:51:04.220456+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmu12s1.kasserver.com","name_value":"nextcloud.kvmu12s1.kasserver.com\nwww.nextcloud.kvmu12s1.kasserver.com","id":10586150052,"entry_timestamp":"2023-09-15T10:36:49.311","not_before":"2023-09-15T09:36:48","not_after":"2023-12-14T09:36:47","serial_number":"045dedd31035da18a73d26fe8a4f963f6eef","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmu12s1.kasserver.com","name_value":"nextcloud.kvmu12s1.kasserver.com\nwww.nextcloud.kvmu12s1.kasserver.com","id":10577792024,"entry_timestamp":"2023-09-15T10:36:49.018","not_before":"2023-09-15T09:36:48","not_after":"2023-12-14T09:36:47","serial_number":"045dedd31035da18a73d26fe8a4f963f6eef","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmu12s1.kasserver.com","name_value":"nextcloud.kvmu12s1.kasserver.com\nwww.nextcloud.kvmu12s1.kasserver.com","id":9948906308,"entry_timestamp":"2023-07-14T08:26:39.958","not_before":"2023-07-14T07:26:39","not_after":"2023-10-12T07:26:38","serial_number":"035e2fe150728c3a351990f06dab0a9b29b7","result_count":3},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"nextcloud.kvmu12s1.kasserver.com","name_value":"nextcloud.kvmu12s1.kasserver.com\nwww.nextcloud.kvmu12s1.kasserver.com","id":9905731790,"entry_timestamp":"2023-07-14T08:26:39.811","not_before":"2023-07-14T07:26:39","not_after":"2023-10-12T07:26:38","serial_number":"035e2fe150728c3a351990f06dab0a9b29b7","result_count":3}]}
|
|
||||||
1
cache/crtsh/kvmwerner2_kasserver_com.json
vendored
1
cache/crtsh/kvmwerner2_kasserver_com.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/llm_mikoshi_de.json
vendored
1
cache/crtsh/llm_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"llm.mikoshi.de","first_cached":"2025-09-15T15:40:09.181904+00:00","last_upstream_query":"2025-09-15T15:40:09.181907+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"llm.mikoshi.de","name_value":"llm.mikoshi.de","id":19995849024,"entry_timestamp":"2025-07-30T00:02:19.725","not_before":"2025-07-29T23:03:47","not_after":"2025-10-27T23:03:46","serial_number":"05335d7df5f076bd039bf6148fe11ebfd86d","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"llm.mikoshi.de","name_value":"llm.mikoshi.de","id":19995848373,"entry_timestamp":"2025-07-30T00:02:17.527","not_before":"2025-07-29T23:03:47","not_after":"2025-10-27T23:03:46","serial_number":"05335d7df5f076bd039bf6148fe11ebfd86d","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"llm.mikoshi.de","name_value":"llm.mikoshi.de","id":18720891589,"entry_timestamp":"2025-05-31T00:01:22.576","not_before":"2025-05-30T23:02:49","not_after":"2025-08-28T23:02:48","serial_number":"064967b3c615cbb4a7f8690b4ee7ed3ab2c2","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"llm.mikoshi.de","name_value":"llm.mikoshi.de","id":18720891502,"entry_timestamp":"2025-05-31T00:01:20.192","not_before":"2025-05-30T23:02:49","not_after":"2025-08-28T23:02:48","serial_number":"064967b3c615cbb4a7f8690b4ee7ed3ab2c2","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"llm.mikoshi.de","name_value":"llm.mikoshi.de","id":17870870943,"entry_timestamp":"2025-03-31T19:43:16.66","not_before":"2025-03-31T18:44:46","not_after":"2025-06-29T18:44:45","serial_number":"05da28c4ebc6a250cb83e37fa4bc54b85508","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"llm.mikoshi.de","name_value":"llm.mikoshi.de","id":17539939653,"entry_timestamp":"2025-03-31T19:43:16.406","not_before":"2025-03-31T18:44:46","not_after":"2025-06-29T18:44:45","serial_number":"05da28c4ebc6a250cb83e37fa4bc54b85508","result_count":2}]}
|
|
||||||
1
cache/crtsh/matrix_cc24_dev.json
vendored
1
cache/crtsh/matrix_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/matrix_mikoshi_de.json
vendored
1
cache/crtsh/matrix_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"matrix.mikoshi.de","first_cached":"2025-09-15T15:40:28.068782+00:00","last_upstream_query":"2025-09-15T15:40:28.068785+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"matrix.mikoshi.de","name_value":"matrix.mikoshi.de","id":17170805427,"entry_timestamp":"2025-02-13T00:01:24.095","not_before":"2025-02-12T23:02:53","not_after":"2025-05-13T23:02:52","serial_number":"0340c3ca26c1ab1678dd4c8885208ac93818","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"matrix.mikoshi.de","name_value":"matrix.mikoshi.de","id":16704107881,"entry_timestamp":"2025-02-13T00:01:23.078","not_before":"2025-02-12T23:02:53","not_after":"2025-05-13T23:02:52","serial_number":"0340c3ca26c1ab1678dd4c8885208ac93818","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"matrix.mikoshi.de","name_value":"matrix.mikoshi.de","id":17122315423,"entry_timestamp":"2025-02-09T15:17:51.214","not_before":"2025-02-09T14:19:20","not_after":"2025-05-10T14:19:19","serial_number":"03e088e499d2b5e10132f28674a00990313a","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"matrix.mikoshi.de","name_value":"matrix.mikoshi.de","id":16635517598,"entry_timestamp":"2025-02-09T15:17:50.818","not_before":"2025-02-09T14:19:20","not_after":"2025-05-10T14:19:19","serial_number":"03e088e499d2b5e10132f28674a00990313a","result_count":2}]}
|
|
||||||
1
cache/crtsh/mempool_mikoshi_de.json
vendored
1
cache/crtsh/mempool_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"mempool.mikoshi.de","first_cached":"2025-09-15T15:37:03.196760+00:00","last_upstream_query":"2025-09-15T15:37:03.196764+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":20275766421,"entry_timestamp":"2025-08-12T00:03:06.545","not_before":"2025-08-11T23:04:33","not_after":"2025-11-09T23:04:32","serial_number":"05db41917aa6146ca29cc3a918e57bfa2dcf","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":20275769957,"entry_timestamp":"2025-08-12T00:03:04.069","not_before":"2025-08-11T23:04:33","not_after":"2025-11-09T23:04:32","serial_number":"05db41917aa6146ca29cc3a918e57bfa2dcf","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":18986859548,"entry_timestamp":"2025-06-13T00:01:16.577","not_before":"2025-06-12T23:02:40","not_after":"2025-09-10T23:02:39","serial_number":"05b7e05db409b0791e0b480cfb8f3116b44f","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":18987057144,"entry_timestamp":"2025-06-13T00:01:10.626","not_before":"2025-06-12T23:02:40","not_after":"2025-09-10T23:02:39","serial_number":"05b7e05db409b0791e0b480cfb8f3116b44f","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":17831239624,"entry_timestamp":"2025-04-14T00:05:10.579","not_before":"2025-04-13T23:06:40","not_after":"2025-07-12T23:06:39","serial_number":"0573b6923c8539ab353225bf023b7fd6b465","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":17831349460,"entry_timestamp":"2025-04-14T00:05:10.362","not_before":"2025-04-13T23:06:40","not_after":"2025-07-12T23:06:39","serial_number":"0573b6923c8539ab353225bf023b7fd6b465","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":17170793138,"entry_timestamp":"2025-02-13T00:00:22.884","not_before":"2025-02-12T23:01:52","not_after":"2025-05-13T23:01:51","serial_number":"048ddc03e0d19d586ef5508bb636fa2ba81b","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":16704095754,"entry_timestamp":"2025-02-13T00:00:22.634","not_before":"2025-02-12T23:01:52","not_after":"2025-05-13T23:01:51","serial_number":"048ddc03e0d19d586ef5508bb636fa2ba81b","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":16636671204,"entry_timestamp":"2025-02-09T15:12:41.443","not_before":"2025-02-09T14:14:11","not_after":"2025-05-10T14:14:10","serial_number":"03a2ba996fd26f676c53edec5725f7442c90","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"mempool.mikoshi.de","name_value":"mempool.mikoshi.de","id":16635433368,"entry_timestamp":"2025-02-09T15:12:41.328","not_before":"2025-02-09T14:14:11","not_after":"2025-05-10T14:14:10","serial_number":"03a2ba996fd26f676c53edec5725f7442c90","result_count":2}]}
|
|
||||||
1
cache/crtsh/mikoshi_de.json
vendored
1
cache/crtsh/mikoshi_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/misp_cc24_dev.json
vendored
1
cache/crtsh/misp_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/misp_mikoshi_de.json
vendored
1
cache/crtsh/misp_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"misp.mikoshi.de","first_cached":"2025-09-14T22:38:25.627275+00:00","last_upstream_query":"2025-09-15T15:37:51.293899+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"misp.mikoshi.de","name_value":"misp.mikoshi.de","id":17353605244,"entry_timestamp":"2025-02-25T00:29:30.385","not_before":"2025-02-24T23:30:59","not_after":"2025-05-25T23:30:58","serial_number":"03b9fa625fe61e6c16d14916613ca1f46d04","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"misp.mikoshi.de","name_value":"misp.mikoshi.de","id":16895667572,"entry_timestamp":"2025-02-25T00:29:29.913","not_before":"2025-02-24T23:30:59","not_after":"2025-05-25T23:30:58","serial_number":"03b9fa625fe61e6c16d14916613ca1f46d04","result_count":2}]}
|
|
||||||
1
cache/crtsh/mx00_ionos_de.json
vendored
1
cache/crtsh/mx00_ionos_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"mx00.ionos.de","first_cached":"2025-09-14T21:05:21.043082+00:00","last_upstream_query":"2025-09-15T15:34:57.508633+00:00","upstream_query_count":3,"certificates":[{"issuer_ca_id":245439,"issuer_name":"C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":17921174855,"entry_timestamp":"2025-04-18T11:32:56.685","not_before":"2024-05-14T10:13:42","not_after":"2025-05-18T23:59:59","serial_number":"01f21195d95cb3f63712c59f40b2f75c","result_count":1},{"issuer_ca_id":245439,"issuer_name":"C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":17755974719,"entry_timestamp":"2025-04-10T06:20:35.546","not_before":"2025-04-10T06:20:33","not_after":"2026-04-14T23:59:59","serial_number":"27efd5b7b17610e4ae86d40dea979ad7","result_count":1},{"issuer_ca_id":245439,"issuer_name":"C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":13038507917,"entry_timestamp":"2024-05-14T10:13:43.64","not_before":"2024-05-14T10:13:42","not_after":"2025-05-18T23:59:59","serial_number":"01f21195d95cb3f63712c59f40b2f75c","result_count":1},{"issuer_ca_id":245439,"issuer_name":"C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":9700366741,"entry_timestamp":"2023-06-20T11:08:14.981","not_before":"2023-06-20T11:08:11","not_after":"2024-06-24T23:59:59","serial_number":"153f3cd769500d3eebec07c90476d817","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":7107715703,"entry_timestamp":"2022-07-12T10:00:03.026","not_before":"2022-07-12T10:00:01","not_after":"2023-07-16T23:59:59","serial_number":"210a3739eb290b28d1199f6a9c04d294","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":4981516894,"entry_timestamp":"2021-08-04T08:35:13.1","not_before":"2021-08-04T08:35:11","not_after":"2022-08-08T23:59:59","serial_number":"11ae5449f5d5cc2a5ec198105748f927","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":2864694015,"entry_timestamp":"2020-05-28T07:54:38.37","not_before":"2020-05-28T07:54:37","not_after":"2022-06-02T23:59:59","serial_number":"325b678601aae53e99926c1834988786","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":2857249553,"entry_timestamp":"2020-05-26T09:01:33.944","not_before":"2020-05-26T09:01:33","not_after":"2022-05-31T23:59:59","serial_number":"0ceee698ba17a744881fbb3998c05748","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":2857204010,"entry_timestamp":"2020-05-26T08:46:52.8","not_before":"2020-05-26T08:46:52","not_after":"2022-05-31T23:59:59","serial_number":"0ac520de92fde640943c1d2c0eb2f6e8","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx00.ionos.de","id":2856876813,"entry_timestamp":"2020-05-26T06:58:49.881","not_before":"2020-05-26T06:58:49","not_after":"2022-05-31T23:59:59","serial_number":"2f9201af62a3408df1579b8a39c0f7b6","result_count":1}]}
|
|
||||||
188
cache/crtsh/mx00_kundenserver_de.json
vendored
188
cache/crtsh/mx00_kundenserver_de.json
vendored
@@ -1,188 +0,0 @@
|
|||||||
{
|
|
||||||
"domain": "mx00.kundenserver.de",
|
|
||||||
"first_cached": "2025-09-15T19:23:52.908544+00:00",
|
|
||||||
"last_upstream_query": "2025-09-15T19:23:52.908548+00:00",
|
|
||||||
"upstream_query_count": 1,
|
|
||||||
"certificates": [
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 245439,
|
|
||||||
"issuer_name": "C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 17921174855,
|
|
||||||
"entry_timestamp": "2025-04-18T11:32:56.685",
|
|
||||||
"not_before": "2024-05-14T10:13:42",
|
|
||||||
"not_after": "2025-05-18T23:59:59",
|
|
||||||
"serial_number": "01f21195d95cb3f63712c59f40b2f75c",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 245439,
|
|
||||||
"issuer_name": "C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 17755974719,
|
|
||||||
"entry_timestamp": "2025-04-10T06:20:35.546",
|
|
||||||
"not_before": "2025-04-10T06:20:33",
|
|
||||||
"not_after": "2026-04-14T23:59:59",
|
|
||||||
"serial_number": "27efd5b7b17610e4ae86d40dea979ad7",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 245439,
|
|
||||||
"issuer_name": "C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 13038507917,
|
|
||||||
"entry_timestamp": "2024-05-14T10:13:43.64",
|
|
||||||
"not_before": "2024-05-14T10:13:42",
|
|
||||||
"not_after": "2025-05-18T23:59:59",
|
|
||||||
"serial_number": "01f21195d95cb3f63712c59f40b2f75c",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 245439,
|
|
||||||
"issuer_name": "C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 9700366741,
|
|
||||||
"entry_timestamp": "2023-06-20T11:08:14.981",
|
|
||||||
"not_before": "2023-06-20T11:08:11",
|
|
||||||
"not_after": "2024-06-24T23:59:59",
|
|
||||||
"serial_number": "153f3cd769500d3eebec07c90476d817",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 6069,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 7107715703,
|
|
||||||
"entry_timestamp": "2022-07-12T10:00:03.026",
|
|
||||||
"not_before": "2022-07-12T10:00:01",
|
|
||||||
"not_after": "2023-07-16T23:59:59",
|
|
||||||
"serial_number": "210a3739eb290b28d1199f6a9c04d294",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 6069,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 4981516894,
|
|
||||||
"entry_timestamp": "2021-08-04T08:35:13.1",
|
|
||||||
"not_before": "2021-08-04T08:35:11",
|
|
||||||
"not_after": "2022-08-08T23:59:59",
|
|
||||||
"serial_number": "11ae5449f5d5cc2a5ec198105748f927",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 6069,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 2864694015,
|
|
||||||
"entry_timestamp": "2020-05-28T07:54:38.37",
|
|
||||||
"not_before": "2020-05-28T07:54:37",
|
|
||||||
"not_after": "2022-06-02T23:59:59",
|
|
||||||
"serial_number": "325b678601aae53e99926c1834988786",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 6069,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 2857249553,
|
|
||||||
"entry_timestamp": "2020-05-26T09:01:33.944",
|
|
||||||
"not_before": "2020-05-26T09:01:33",
|
|
||||||
"not_after": "2022-05-31T23:59:59",
|
|
||||||
"serial_number": "0ceee698ba17a744881fbb3998c05748",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 6069,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 2857204010,
|
|
||||||
"entry_timestamp": "2020-05-26T08:46:52.8",
|
|
||||||
"not_before": "2020-05-26T08:46:52",
|
|
||||||
"not_after": "2022-05-31T23:59:59",
|
|
||||||
"serial_number": "0ac520de92fde640943c1d2c0eb2f6e8",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 6069,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 2856876813,
|
|
||||||
"entry_timestamp": "2020-05-26T06:58:49.881",
|
|
||||||
"not_before": "2020-05-26T06:58:49",
|
|
||||||
"not_after": "2022-05-31T23:59:59",
|
|
||||||
"serial_number": "2f9201af62a3408df1579b8a39c0f7b6",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 6069,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 1351991065,
|
|
||||||
"entry_timestamp": "2019-04-05T05:26:06.601",
|
|
||||||
"not_before": "2019-02-18T09:14:39",
|
|
||||||
"not_after": "2021-02-23T23:59:59",
|
|
||||||
"serial_number": "2c12501e6e490dd60bfb87d85edad1d1",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 6069,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 1213296236,
|
|
||||||
"entry_timestamp": "2019-02-18T09:14:39.61",
|
|
||||||
"not_before": "2019-02-18T09:14:39",
|
|
||||||
"not_after": "2021-02-23T23:59:59",
|
|
||||||
"serial_number": "2c12501e6e490dd60bfb87d85edad1d1",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 1375,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=NRW, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass DE-1",
|
|
||||||
"common_name": "mx00.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 370285812,
|
|
||||||
"entry_timestamp": "2018-03-30T01:09:17.252",
|
|
||||||
"not_before": "2013-08-06T15:49:56",
|
|
||||||
"not_after": "2014-08-11T23:59:59",
|
|
||||||
"serial_number": "00f17c74ae3d6e7a5c",
|
|
||||||
"result_count": 2
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 1477,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass DE-2",
|
|
||||||
"common_name": "mx.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 307747879,
|
|
||||||
"entry_timestamp": "2018-01-19T09:01:19.72",
|
|
||||||
"not_before": "2017-03-22T10:21:38",
|
|
||||||
"not_after": "2019-03-27T23:59:59",
|
|
||||||
"serial_number": "00bf95dcc292d7bf01",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 1477,
|
|
||||||
"issuer_name": "C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass DE-2",
|
|
||||||
"common_name": "mx00.kundenserver.de",
|
|
||||||
"name_value": "mx00.kundenserver.de",
|
|
||||||
"id": 14627870,
|
|
||||||
"entry_timestamp": "2016-03-08T11:54:51.83",
|
|
||||||
"not_before": "2014-07-22T11:15:15",
|
|
||||||
"not_after": "2017-07-27T23:59:59",
|
|
||||||
"serial_number": "20e14d5c7f18e199",
|
|
||||||
"result_count": 2
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
1
cache/crtsh/mx01_ionos_de.json
vendored
1
cache/crtsh/mx01_ionos_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"mx01.ionos.de","first_cached":"2025-09-14T21:05:58.567956+00:00","last_upstream_query":"2025-09-15T15:34:59.570960+00:00","upstream_query_count":3,"certificates":[{"issuer_ca_id":245439,"issuer_name":"C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":17921174855,"entry_timestamp":"2025-04-18T11:32:56.685","not_before":"2024-05-14T10:13:42","not_after":"2025-05-18T23:59:59","serial_number":"01f21195d95cb3f63712c59f40b2f75c","result_count":1},{"issuer_ca_id":245439,"issuer_name":"C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":17755974719,"entry_timestamp":"2025-04-10T06:20:35.546","not_before":"2025-04-10T06:20:33","not_after":"2026-04-14T23:59:59","serial_number":"27efd5b7b17610e4ae86d40dea979ad7","result_count":1},{"issuer_ca_id":245439,"issuer_name":"C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":13038507917,"entry_timestamp":"2024-05-14T10:13:43.64","not_before":"2024-05-14T10:13:42","not_after":"2025-05-18T23:59:59","serial_number":"01f21195d95cb3f63712c59f40b2f75c","result_count":1},{"issuer_ca_id":245439,"issuer_name":"C=DE, O=Deutsche Telekom Security GmbH, CN=Telekom Security ServerID OV Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":9700366741,"entry_timestamp":"2023-06-20T11:08:14.981","not_before":"2023-06-20T11:08:11","not_after":"2024-06-24T23:59:59","serial_number":"153f3cd769500d3eebec07c90476d817","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":7107715703,"entry_timestamp":"2022-07-12T10:00:03.026","not_before":"2022-07-12T10:00:01","not_after":"2023-07-16T23:59:59","serial_number":"210a3739eb290b28d1199f6a9c04d294","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":4981516894,"entry_timestamp":"2021-08-04T08:35:13.1","not_before":"2021-08-04T08:35:11","not_after":"2022-08-08T23:59:59","serial_number":"11ae5449f5d5cc2a5ec198105748f927","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":2864694015,"entry_timestamp":"2020-05-28T07:54:38.37","not_before":"2020-05-28T07:54:37","not_after":"2022-06-02T23:59:59","serial_number":"325b678601aae53e99926c1834988786","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":2857249553,"entry_timestamp":"2020-05-26T09:01:33.944","not_before":"2020-05-26T09:01:33","not_after":"2022-05-31T23:59:59","serial_number":"0ceee698ba17a744881fbb3998c05748","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":2857204010,"entry_timestamp":"2020-05-26T08:46:52.8","not_before":"2020-05-26T08:46:52","not_after":"2022-05-31T23:59:59","serial_number":"0ac520de92fde640943c1d2c0eb2f6e8","result_count":1},{"issuer_ca_id":6069,"issuer_name":"C=DE, O=T-Systems International GmbH, OU=T-Systems Trust Center, ST=Nordrhein Westfalen, postalCode=57250, L=Netphen, street=Untere Industriestr. 20, CN=TeleSec ServerPass Class 2 CA","common_name":"mx.kundenserver.de","name_value":"mx01.ionos.de","id":2856876813,"entry_timestamp":"2020-05-26T06:58:49.881","not_before":"2020-05-26T06:58:49","not_after":"2022-05-31T23:59:59","serial_number":"2f9201af62a3408df1579b8a39c0f7b6","result_count":1}]}
|
|
||||||
1
cache/crtsh/mx01_kundenserver_de.json
vendored
1
cache/crtsh/mx01_kundenserver_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/mx_kundenserver_de.json
vendored
1
cache/crtsh/mx_kundenserver_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/nostr_mikoshi_de.json
vendored
1
cache/crtsh/nostr_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"nostr.mikoshi.de","first_cached":"2025-09-14T22:44:55.522098+00:00","last_upstream_query":"2025-09-15T15:39:39.928858+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295809,"issuer_name":"C=US, O=Let's Encrypt, CN=E8","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":20839972270,"entry_timestamp":"2025-09-07T00:01:25.998","not_before":"2025-09-06T23:02:55","not_after":"2025-12-05T23:02:54","serial_number":"05c5c3ab0353acb84d7adc05e78d6d6fe276","result_count":2},{"issuer_ca_id":295809,"issuer_name":"C=US, O=Let's Encrypt, CN=E8","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":20839069128,"entry_timestamp":"2025-09-07T00:01:25.786","not_before":"2025-09-06T23:02:55","not_after":"2025-12-05T23:02:54","serial_number":"05c5c3ab0353acb84d7adc05e78d6d6fe276","result_count":2},{"issuer_ca_id":295813,"issuer_name":"C=US, O=Let's Encrypt, CN=E7","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":20641626677,"entry_timestamp":"2025-08-29T00:03:39.504","not_before":"2025-08-28T23:05:09","not_after":"2025-11-26T23:05:08","serial_number":"059eab3e56a39c431f4cb961af259f45f550","result_count":2},{"issuer_ca_id":295813,"issuer_name":"C=US, O=Let's Encrypt, CN=E7","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":20641627201,"entry_timestamp":"2025-08-29T00:03:39.306","not_before":"2025-08-28T23:05:09","not_after":"2025-11-26T23:05:08","serial_number":"059eab3e56a39c431f4cb961af259f45f550","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":20446505609,"entry_timestamp":"2025-08-19T20:58:17.951","not_before":"2025-08-19T19:59:45","not_after":"2025-11-17T19:59:44","serial_number":"06e00b2fda8f18091bb0be67f6809f87a889","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":20441143198,"entry_timestamp":"2025-08-19T20:58:15.697","not_before":"2025-08-19T19:59:45","not_after":"2025-11-17T19:59:44","serial_number":"06e00b2fda8f18091bb0be67f6809f87a889","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":17170798515,"entry_timestamp":"2025-02-13T00:00:50.437","not_before":"2025-02-12T23:02:19","not_after":"2025-05-13T23:02:18","serial_number":"04b641da285b3701db3e199cc49b87a023c7","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":16705237488,"entry_timestamp":"2025-02-13T00:00:50.226","not_before":"2025-02-12T23:02:19","not_after":"2025-05-13T23:02:18","serial_number":"04b641da285b3701db3e199cc49b87a023c7","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":17122272223,"entry_timestamp":"2025-02-09T15:15:01.538","not_before":"2025-02-09T14:16:30","not_after":"2025-05-10T14:16:29","serial_number":"04f78d3955d454fb6d004d08cfadb198d2be","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"nostr.mikoshi.de","name_value":"nostr.mikoshi.de","id":16636445173,"entry_timestamp":"2025-02-09T15:15:01.221","not_before":"2025-02-09T14:16:30","not_after":"2025-05-10T14:16:29","serial_number":"04f78d3955d454fb6d004d08cfadb198d2be","result_count":2}]}
|
|
||||||
1
cache/crtsh/notes_mikoshi_de.json
vendored
1
cache/crtsh/notes_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"notes.mikoshi.de","first_cached":"2025-09-15T15:37:27.243670+00:00","last_upstream_query":"2025-09-15T15:37:27.243675+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"notes.mikoshi.de","name_value":"notes.mikoshi.de","id":17831397265,"entry_timestamp":"2025-04-14T00:15:27.018","not_before":"2025-04-13T23:16:56","not_after":"2025-07-12T23:16:55","serial_number":"0604f6b0149b4afe45efbe46a5493c1ec3fd","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"notes.mikoshi.de","name_value":"notes.mikoshi.de","id":17831343922,"entry_timestamp":"2025-04-14T00:15:26.683","not_before":"2025-04-13T23:16:56","not_after":"2025-07-12T23:16:55","serial_number":"0604f6b0149b4afe45efbe46a5493c1ec3fd","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"notes.mikoshi.de","name_value":"notes.mikoshi.de","id":17170808839,"entry_timestamp":"2025-02-13T00:01:30.661","not_before":"2025-02-12T23:03:00","not_after":"2025-05-13T23:02:59","serial_number":"032b2779b00fd8c25fb56ccab4a5eccf7b1a","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"notes.mikoshi.de","name_value":"notes.mikoshi.de","id":16704108124,"entry_timestamp":"2025-02-13T00:01:30.468","not_before":"2025-02-12T23:03:00","not_after":"2025-05-13T23:02:59","serial_number":"032b2779b00fd8c25fb56ccab4a5eccf7b1a","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"notes.mikoshi.de","name_value":"notes.mikoshi.de","id":17122334548,"entry_timestamp":"2025-02-09T15:18:27.206","not_before":"2025-02-09T14:19:56","not_after":"2025-05-10T14:19:55","serial_number":"042257f7bcc91c0994e9c3c76da6ee700a03","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"notes.mikoshi.de","name_value":"notes.mikoshi.de","id":16636453516,"entry_timestamp":"2025-02-09T15:18:26.537","not_before":"2025-02-09T14:19:56","not_after":"2025-05-10T14:19:55","serial_number":"042257f7bcc91c0994e9c3c76da6ee700a03","result_count":2}]}
|
|
||||||
1
cache/crtsh/ns3_m-online_net.json
vendored
1
cache/crtsh/ns3_m-online_net.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/ns4_m-online_net.json
vendored
1
cache/crtsh/ns4_m-online_net.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/office_mikoshi_de.json
vendored
1
cache/crtsh/office_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"office.mikoshi.de","first_cached":"2025-09-15T15:38:27.549467+00:00","last_upstream_query":"2025-09-15T15:38:27.549473+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":20275820856,"entry_timestamp":"2025-08-12T00:07:33.018","not_before":"2025-08-11T23:09:02","not_after":"2025-11-09T23:09:01","serial_number":"05c2c936c2f12251f76b59d3ad2854f50f78","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":20275819318,"entry_timestamp":"2025-08-12T00:07:32.912","not_before":"2025-08-11T23:09:02","not_after":"2025-11-09T23:09:01","serial_number":"05c2c936c2f12251f76b59d3ad2854f50f78","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":18987110425,"entry_timestamp":"2025-06-13T00:04:44.238","not_before":"2025-06-12T23:06:13","not_after":"2025-09-10T23:06:12","serial_number":"051acaa08d5d89e6f4063edabca209961196","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":18987111139,"entry_timestamp":"2025-06-13T00:04:43.915","not_before":"2025-06-12T23:06:13","not_after":"2025-09-10T23:06:12","serial_number":"051acaa08d5d89e6f4063edabca209961196","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":17831462745,"entry_timestamp":"2025-04-14T00:16:05.147","not_before":"2025-04-13T23:17:32","not_after":"2025-07-12T23:17:31","serial_number":"054ddaf61e4b9f57e8d56a9497df9149c395","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":17831369811,"entry_timestamp":"2025-04-14T00:16:03.009","not_before":"2025-04-13T23:17:32","not_after":"2025-07-12T23:17:31","serial_number":"054ddaf61e4b9f57e8d56a9497df9149c395","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":17170808669,"entry_timestamp":"2025-02-13T00:01:46.473","not_before":"2025-02-12T23:03:15","not_after":"2025-05-13T23:03:14","serial_number":"03deb0895b1825adbbbaf9addcd01dd92799","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":16720296902,"entry_timestamp":"2025-02-13T00:01:45.743","not_before":"2025-02-12T23:03:15","not_after":"2025-05-13T23:03:14","serial_number":"03deb0895b1825adbbbaf9addcd01dd92799","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":17122359365,"entry_timestamp":"2025-02-09T15:19:36.715","not_before":"2025-02-09T14:21:06","not_after":"2025-05-10T14:21:05","serial_number":"04610182dee49a266a3d43aadacfd0b41e27","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"office.mikoshi.de","name_value":"office.mikoshi.de","id":16635528699,"entry_timestamp":"2025-02-09T15:19:36.337","not_before":"2025-02-09T14:21:06","not_after":"2025-05-10T14:21:05","serial_number":"04610182dee49a266a3d43aadacfd0b41e27","result_count":2}]}
|
|
||||||
1
cache/crtsh/overcuriousity_org.json
vendored
1
cache/crtsh/overcuriousity_org.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/piped_mikoshi_de.json
vendored
1
cache/crtsh/piped_mikoshi_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/proxy_piped_mikoshi_de.json
vendored
1
cache/crtsh/proxy_piped_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"proxy.piped.mikoshi.de","first_cached":"2025-09-14T22:45:40.325331+00:00","last_upstream_query":"2025-09-15T15:38:51.713680+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":20189220475,"entry_timestamp":"2025-08-08T00:05:07.009","not_before":"2025-08-07T23:06:36","not_after":"2025-11-05T23:06:35","serial_number":"05c0d9d7f7cae49c673abcdb28c4412d78d6","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":20189220702,"entry_timestamp":"2025-08-08T00:05:06.824","not_before":"2025-08-07T23:06:36","not_after":"2025-11-05T23:06:35","serial_number":"05c0d9d7f7cae49c673abcdb28c4412d78d6","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":20060639877,"entry_timestamp":"2025-08-02T00:01:52.165","not_before":"2025-08-01T23:03:21","not_after":"2025-10-30T23:03:20","serial_number":"05cb1fd70bdfe7e07c8a952b4e089c39d1da","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":20060638194,"entry_timestamp":"2025-08-02T00:01:51.846","not_before":"2025-08-01T23:03:21","not_after":"2025-10-30T23:03:20","serial_number":"05cb1fd70bdfe7e07c8a952b4e089c39d1da","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":19929369417,"entry_timestamp":"2025-07-27T00:02:26.4","not_before":"2025-07-26T23:03:54","not_after":"2025-10-24T23:03:53","serial_number":"063f8984c7c967816e258a7216ecd1959f34","result_count":2},{"issuer_ca_id":295810,"issuer_name":"C=US, O=Let's Encrypt, CN=E5","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":19929369596,"entry_timestamp":"2025-07-27T00:02:24.492","not_before":"2025-07-26T23:03:54","not_after":"2025-10-24T23:03:53","serial_number":"063f8984c7c967816e258a7216ecd1959f34","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":19803763772,"entry_timestamp":"2025-07-21T00:00:57.2","not_before":"2025-07-20T23:02:26","not_after":"2025-10-18T23:02:25","serial_number":"0572a8c3609d51428351eb07244bcd441072","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":19803763758,"entry_timestamp":"2025-07-21T00:00:56.856","not_before":"2025-07-20T23:02:26","not_after":"2025-10-18T23:02:25","serial_number":"0572a8c3609d51428351eb07244bcd441072","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":19668416256,"entry_timestamp":"2025-07-15T00:00:58.486","not_before":"2025-07-14T23:02:28","not_after":"2025-10-12T23:02:27","serial_number":"06664191409b60fb8f1241b999cd884247cb","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":19668412981,"entry_timestamp":"2025-07-15T00:00:58.222","not_before":"2025-07-14T23:02:28","not_after":"2025-10-12T23:02:27","serial_number":"06664191409b60fb8f1241b999cd884247cb","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":19523857935,"entry_timestamp":"2025-07-08T06:31:01.3","not_before":"2025-07-08T05:32:26","not_after":"2025-10-06T05:32:25","serial_number":"05def41b49262bc3c90ceef09a3c53fdb897","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"proxy.piped.mikoshi.de","name_value":"proxy.piped.mikoshi.de","id":19523855448,"entry_timestamp":"2025-07-08T06:30:56.952","not_before":"2025-07-08T05:32:26","not_after":"2025-10-06T05:32:25","serial_number":"05def41b49262bc3c90ceef09a3c53fdb897","result_count":2}]}
|
|
||||||
1
cache/crtsh/push_mikoshi_de.json
vendored
1
cache/crtsh/push_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"push.mikoshi.de","first_cached":"2025-09-15T15:36:51.090852+00:00","last_upstream_query":"2025-09-15T15:36:51.090855+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":20275820653,"entry_timestamp":"2025-08-12T00:06:47.342","not_before":"2025-08-11T23:08:16","not_after":"2025-11-09T23:08:15","serial_number":"052026f4f3142d63bd9ca3feae317ece4ee9","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":20275820169,"entry_timestamp":"2025-08-12T00:06:47.109","not_before":"2025-08-11T23:08:16","not_after":"2025-11-09T23:08:15","serial_number":"052026f4f3142d63bd9ca3feae317ece4ee9","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":18987077850,"entry_timestamp":"2025-06-13T00:04:10.192","not_before":"2025-06-12T23:05:37","not_after":"2025-09-10T23:05:36","serial_number":"0546092d0c32d96eb7fcf0ce30e5e65ea3e1","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":18987099791,"entry_timestamp":"2025-06-13T00:04:07.918","not_before":"2025-06-12T23:05:37","not_after":"2025-09-10T23:05:36","serial_number":"0546092d0c32d96eb7fcf0ce30e5e65ea3e1","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":17831474873,"entry_timestamp":"2025-04-14T00:14:50.26","not_before":"2025-04-13T23:16:19","not_after":"2025-07-12T23:16:18","serial_number":"0513fd56736b1233441a3275762b03fc9eb0","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":17831474384,"entry_timestamp":"2025-04-14T00:14:50.004","not_before":"2025-04-13T23:16:19","not_after":"2025-07-12T23:16:18","serial_number":"0513fd56736b1233441a3275762b03fc9eb0","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":17170801795,"entry_timestamp":"2025-02-13T00:01:17.002","not_before":"2025-02-12T23:02:46","not_after":"2025-05-13T23:02:45","serial_number":"03487570ab1eaa62f8f4c3b61a8b90eda732","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":16715026255,"entry_timestamp":"2025-02-13T00:01:16.729","not_before":"2025-02-12T23:02:46","not_after":"2025-05-13T23:02:45","serial_number":"03487570ab1eaa62f8f4c3b61a8b90eda732","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":17122309940,"entry_timestamp":"2025-02-09T15:17:17.076","not_before":"2025-02-09T14:18:46","not_after":"2025-05-10T14:18:45","serial_number":"047ea82614cb4449bda99fc1ba86eb49b5c8","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"push.mikoshi.de","name_value":"push.mikoshi.de","id":16635500733,"entry_timestamp":"2025-02-09T15:17:16.594","not_before":"2025-02-09T14:18:46","not_after":"2025-05-10T14:18:45","serial_number":"047ea82614cb4449bda99fc1ba86eb49b5c8","result_count":2}]}
|
|
||||||
1
cache/crtsh/raptor_cc24_dev.json
vendored
1
cache/crtsh/raptor_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/s3_cc24_dev.json
vendored
1
cache/crtsh/s3_cc24_dev.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"s3.cc24.dev","first_cached":"2025-09-14T21:38:35.568153+00:00","last_upstream_query":"2025-09-14T21:38:35.568155+00:00","upstream_query_count":1,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"console.s3.cc24.dev","name_value":"console.s3.cc24.dev","id":20287575466,"entry_timestamp":"2025-08-12T12:55:37.077","not_before":"2025-08-12T11:57:05","not_after":"2025-11-10T11:57:04","serial_number":"066bdfa83088f8d7e67284da94dd5d122ed6","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"console.s3.cc24.dev","name_value":"console.s3.cc24.dev","id":20287575457,"entry_timestamp":"2025-08-12T12:55:36.75","not_before":"2025-08-12T11:57:05","not_after":"2025-11-10T11:57:04","serial_number":"066bdfa83088f8d7e67284da94dd5d122ed6","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"s3.cc24.dev","name_value":"s3.cc24.dev","id":20285693027,"entry_timestamp":"2025-08-12T10:58:41.611","not_before":"2025-08-12T10:00:11","not_after":"2025-11-10T10:00:10","serial_number":"06f0f73404258136977fdbfe4cf51db786a7","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"s3.cc24.dev","name_value":"s3.cc24.dev","id":20285693495,"entry_timestamp":"2025-08-12T10:58:41.366","not_before":"2025-08-12T10:00:11","not_after":"2025-11-10T10:00:10","serial_number":"06f0f73404258136977fdbfe4cf51db786a7","result_count":2}]}
|
|
||||||
1
cache/crtsh/se_mikoshi_de.json
vendored
1
cache/crtsh/se_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"se.mikoshi.de","first_cached":"2025-09-14T22:37:29.207379+00:00","last_upstream_query":"2025-09-15T15:40:52.328157+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":20296592765,"entry_timestamp":"2025-08-13T00:02:43.765","not_before":"2025-08-12T23:04:13","not_after":"2025-11-10T23:04:12","serial_number":"05688b90444fda4afff5208583d9b4c6b453","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":20296592676,"entry_timestamp":"2025-08-13T00:02:43.253","not_before":"2025-08-12T23:04:13","not_after":"2025-11-10T23:04:12","serial_number":"05688b90444fda4afff5208583d9b4c6b453","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":19007481148,"entry_timestamp":"2025-06-14T00:01:17.383","not_before":"2025-06-13T23:02:46","not_after":"2025-09-11T23:02:45","serial_number":"057f456b58a7ea34177d50585ea3b4561e6b","result_count":2},{"issuer_ca_id":295814,"issuer_name":"C=US, O=Let's Encrypt, CN=R10","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":19007493221,"entry_timestamp":"2025-06-14T00:01:17.049","not_before":"2025-06-13T23:02:46","not_after":"2025-09-11T23:02:45","serial_number":"057f456b58a7ea34177d50585ea3b4561e6b","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":17850961291,"entry_timestamp":"2025-04-15T00:00:37.293","not_before":"2025-04-14T23:02:02","not_after":"2025-07-13T23:02:01","serial_number":"052c5b098d5fa135547d9e3d7f7496b63af6","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":17850962036,"entry_timestamp":"2025-04-15T00:00:32.962","not_before":"2025-04-14T23:02:02","not_after":"2025-07-13T23:02:01","serial_number":"052c5b098d5fa135547d9e3d7f7496b63af6","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":17170796076,"entry_timestamp":"2025-02-13T00:00:29.982","not_before":"2025-02-12T23:01:59","not_after":"2025-05-13T23:01:58","serial_number":"04ecdefe30d1b5ce4d3ebb91bbbb2257bd15","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":16704095774,"entry_timestamp":"2025-02-13T00:00:29.674","not_before":"2025-02-12T23:01:59","not_after":"2025-05-13T23:01:58","serial_number":"04ecdefe30d1b5ce4d3ebb91bbbb2257bd15","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":17122261002,"entry_timestamp":"2025-02-09T15:13:15.78","not_before":"2025-02-09T14:14:45","not_after":"2025-05-10T14:14:44","serial_number":"03972d789d00e59045e4f37eae917438e10a","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":16635473126,"entry_timestamp":"2025-02-09T15:13:15.412","not_before":"2025-02-09T14:14:45","not_after":"2025-05-10T14:14:44","serial_number":"03972d789d00e59045e4f37eae917438e10a","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":11490113222,"entry_timestamp":"2023-12-16T00:55:58.472","not_before":"2023-12-15T23:55:57","not_after":"2024-03-14T23:55:56","serial_number":"0386fd7fd921ad63f4805ad4f593819aa238","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"se.mikoshi.de","name_value":"se.mikoshi.de","id":11422214995,"entry_timestamp":"2023-12-16T00:55:57.647","not_before":"2023-12-15T23:55:57","not_after":"2024-03-14T23:55:56","serial_number":"0386fd7fd921ad63f4805ad4f593819aa238","result_count":2}]}
|
|
||||||
1
cache/crtsh/signaling_mikoshi_de.json
vendored
1
cache/crtsh/signaling_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"signaling.mikoshi.de","first_cached":"2025-09-14T21:05:14.189157+00:00","last_upstream_query":"2025-09-15T15:36:14.904100+00:00","upstream_query_count":3,"certificates":[{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":19208276669,"entry_timestamp":"2025-06-23T18:21:11.885","not_before":"2025-06-23T17:22:40","not_after":"2025-09-21T17:22:39","serial_number":"0542fce9cb99bb1c1d18e5e452c90d850936","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":19208276557,"entry_timestamp":"2025-06-23T18:21:11.215","not_before":"2025-06-23T17:22:40","not_after":"2025-09-21T17:22:39","serial_number":"0542fce9cb99bb1c1d18e5e452c90d850936","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":19208272013,"entry_timestamp":"2025-06-23T18:20:29.619","not_before":"2025-06-23T17:21:56","not_after":"2025-09-21T17:21:55","serial_number":"0540be6c8cb99dcaa5492af7b934f40466f9","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":19208263539,"entry_timestamp":"2025-06-23T18:20:26.82","not_before":"2025-06-23T17:21:56","not_after":"2025-09-21T17:21:55","serial_number":"0540be6c8cb99dcaa5492af7b934f40466f9","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":17428224909,"entry_timestamp":"2025-03-23T15:21:03.771","not_before":"2025-03-23T14:22:33","not_after":"2025-06-21T14:22:32","serial_number":"0537513d69487cead5f018b5322aa54fb52e","result_count":2},{"issuer_ca_id":295819,"issuer_name":"C=US, O=Let's Encrypt, CN=E6","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":17346408492,"entry_timestamp":"2025-03-23T15:21:03.617","not_before":"2025-03-23T14:22:33","not_after":"2025-06-21T14:22:32","serial_number":"0537513d69487cead5f018b5322aa54fb52e","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":17170806575,"entry_timestamp":"2025-02-13T00:01:39.507","not_before":"2025-02-12T23:03:07","not_after":"2025-05-13T23:03:06","serial_number":"04a3141d80551d2287647211445f32131b19","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":16704105641,"entry_timestamp":"2025-02-13T00:01:37.463","not_before":"2025-02-12T23:03:07","not_after":"2025-05-13T23:03:06","serial_number":"04a3141d80551d2287647211445f32131b19","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":17122350772,"entry_timestamp":"2025-02-09T15:19:01.645","not_before":"2025-02-09T14:20:31","not_after":"2025-05-10T14:20:30","serial_number":"0456c53707eac4c8e79109b09eb9257674ef","result_count":2},{"issuer_ca_id":295815,"issuer_name":"C=US, O=Let's Encrypt, CN=R11","common_name":"signaling.mikoshi.de","name_value":"signaling.mikoshi.de","id":16635514012,"entry_timestamp":"2025-02-09T15:19:01.438","not_before":"2025-02-09T14:20:31","not_after":"2025-05-10T14:20:30","serial_number":"0456c53707eac4c8e79109b09eb9257674ef","result_count":2}]}
|
|
||||||
1
cache/crtsh/status_mikoshi_de.json
vendored
1
cache/crtsh/status_mikoshi_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/streaming_mikoshi_de.json
vendored
1
cache/crtsh/streaming_mikoshi_de.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/test_mikoshi_de.json
vendored
1
cache/crtsh/test_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"test.mikoshi.de","first_cached":"2025-09-14T22:38:58.849950+00:00","last_upstream_query":"2025-09-15T15:34:55.551888+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"test.mikoshi.de","name_value":"test.mikoshi.de","id":11527598498,"entry_timestamp":"2023-12-19T19:30:11.511","not_before":"2023-12-19T18:30:10","not_after":"2024-03-18T18:30:09","serial_number":"0460a2ce513e2971556b911e64d3c06c876d","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"test.mikoshi.de","name_value":"test.mikoshi.de","id":11448041639,"entry_timestamp":"2023-12-19T19:30:10.729","not_before":"2023-12-19T18:30:10","not_after":"2024-03-18T18:30:09","serial_number":"0460a2ce513e2971556b911e64d3c06c876d","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"test.mikoshi.de","name_value":"test.mikoshi.de","id":11487822748,"entry_timestamp":"2023-12-15T22:27:24.154","not_before":"2023-12-15T21:27:23","not_after":"2024-03-14T21:27:22","serial_number":"03524dd265a8db8b1d3b8ae9236457b06564","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"test.mikoshi.de","name_value":"test.mikoshi.de","id":11420685419,"entry_timestamp":"2023-12-15T22:27:23.743","not_before":"2023-12-15T21:27:23","not_after":"2024-03-14T21:27:22","serial_number":"03524dd265a8db8b1d3b8ae9236457b06564","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"test.mikoshi.de","name_value":"test.mikoshi.de","id":10845362810,"entry_timestamp":"2023-10-20T20:03:11.855","not_before":"2023-10-20T19:03:11","not_after":"2024-01-18T19:03:10","serial_number":"04fb05ec64b2614143f419dfe3c1e058a6f8","result_count":2},{"issuer_ca_id":183267,"issuer_name":"C=US, O=Let's Encrypt, CN=R3","common_name":"test.mikoshi.de","name_value":"test.mikoshi.de","id":10845356148,"entry_timestamp":"2023-10-20T20:03:11.585","not_before":"2023-10-20T19:03:11","not_after":"2024-01-18T19:03:10","serial_number":"04fb05ec64b2614143f419dfe3c1e058a6f8","result_count":2}]}
|
|
||||||
1
cache/crtsh/timesketch_cc24_dev.json
vendored
1
cache/crtsh/timesketch_cc24_dev.json
vendored
File diff suppressed because one or more lines are too long
1
cache/crtsh/vm0800_kasserver_com.json
vendored
1
cache/crtsh/vm0800_kasserver_com.json
vendored
File diff suppressed because one or more lines are too long
32
cache/crtsh/www_overcuriousity_org.json
vendored
32
cache/crtsh/www_overcuriousity_org.json
vendored
@@ -1,32 +0,0 @@
|
|||||||
{
|
|
||||||
"domain": "www.overcuriousity.org",
|
|
||||||
"first_cached": "2025-09-14T21:16:24.041839+00:00",
|
|
||||||
"last_upstream_query": "2025-09-15T19:08:25.160183+00:00",
|
|
||||||
"upstream_query_count": 3,
|
|
||||||
"certificates": [
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 295819,
|
|
||||||
"issuer_name": "C=US, O=Let's Encrypt, CN=E6",
|
|
||||||
"common_name": "signaling.mikoshi.de",
|
|
||||||
"name_value": "www.overcuriousity.org",
|
|
||||||
"id": 19208272013,
|
|
||||||
"entry_timestamp": "2025-06-23T18:20:29.619",
|
|
||||||
"not_before": "2025-06-23T17:21:56",
|
|
||||||
"not_after": "2025-09-21T17:21:55",
|
|
||||||
"serial_number": "0540be6c8cb99dcaa5492af7b934f40466f9",
|
|
||||||
"result_count": 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"issuer_ca_id": 295819,
|
|
||||||
"issuer_name": "C=US, O=Let's Encrypt, CN=E6",
|
|
||||||
"common_name": "signaling.mikoshi.de",
|
|
||||||
"name_value": "www.overcuriousity.org",
|
|
||||||
"id": 19208263539,
|
|
||||||
"entry_timestamp": "2025-06-23T18:20:26.82",
|
|
||||||
"not_before": "2025-06-23T17:21:56",
|
|
||||||
"not_after": "2025-09-21T17:21:55",
|
|
||||||
"serial_number": "0540be6c8cb99dcaa5492af7b934f40466f9",
|
|
||||||
"result_count": 1
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
1
cache/crtsh/zap_mikoshi_de.json
vendored
1
cache/crtsh/zap_mikoshi_de.json
vendored
@@ -1 +0,0 @@
|
|||||||
{"domain":"zap.mikoshi.de","first_cached":"2025-09-14T22:37:56.901178+00:00","last_upstream_query":"2025-09-15T15:36:02.901638+00:00","upstream_query_count":2,"certificates":[{"issuer_ca_id":295813,"issuer_name":"C=US, O=Let's Encrypt, CN=E7","common_name":"zap.mikoshi.de","name_value":"zap.mikoshi.de","id":20979146330,"entry_timestamp":"2025-09-13T00:01:36.857","not_before":"2025-09-12T23:03:06","not_after":"2025-12-11T23:03:05","serial_number":"060330123144270f9c8614ee282e43b5b676","result_count":2},{"issuer_ca_id":295813,"issuer_name":"C=US, O=Let's Encrypt, CN=E7","common_name":"zap.mikoshi.de","name_value":"zap.mikoshi.de","id":20979146321,"entry_timestamp":"2025-09-13T00:01:36.705","not_before":"2025-09-12T23:03:06","not_after":"2025-12-11T23:03:05","serial_number":"060330123144270f9c8614ee282e43b5b676","result_count":2},{"issuer_ca_id":295809,"issuer_name":"C=US, O=Let's Encrypt, CN=E8","common_name":"zap.mikoshi.de","name_value":"zap.mikoshi.de","id":20955786867,"entry_timestamp":"2025-09-12T00:00:06.78","not_before":"2025-09-11T23:01:36","not_after":"2025-12-10T23:01:35","serial_number":"069fdbfcf2023341872c23d4eadf03dbd600","result_count":2},{"issuer_ca_id":295809,"issuer_name":"C=US, O=Let's Encrypt, CN=E8","common_name":"zap.mikoshi.de","name_value":"zap.mikoshi.de","id":20955786725,"entry_timestamp":"2025-09-12T00:00:06.615","not_before":"2025-09-11T23:01:36","not_after":"2025-12-10T23:01:35","serial_number":"069fdbfcf2023341872c23d4eadf03dbd600","result_count":2},{"issuer_ca_id":295809,"issuer_name":"C=US, O=Let's Encrypt, CN=E8","common_name":"zap.mikoshi.de","name_value":"zap.mikoshi.de","id":20754243960,"entry_timestamp":"2025-09-03T00:01:13.766","not_before":"2025-09-02T23:02:41","not_after":"2025-12-01T23:02:40","serial_number":"05c8fbc52b5d8c0ca2089679fd1574503cf0","result_count":2},{"issuer_ca_id":295809,"issuer_name":"C=US, O=Let's Encrypt, CN=E8","common_name":"zap.mikoshi.de","name_value":"zap.mikoshi.de","id":20754243261,"entry_timestamp":"2025-09-03T00:01:11.505","not_before":"2025-09-02T23:02:41","not_after":"2025-12-01T23:02:40","serial_number":"05c8fbc52b5d8c0ca2089679fd1574503cf0","result_count":2},{"issuer_ca_id":295813,"issuer_name":"C=US, O=Let's Encrypt, CN=E7","common_name":"zap.mikoshi.de","name_value":"zap.mikoshi.de","id":20539947210,"entry_timestamp":"2025-08-24T10:34:58.674","not_before":"2025-08-24T09:36:27","not_after":"2025-11-22T09:36:26","serial_number":"062deaf87750f6cf54f0027bbf6f9617447b","result_count":2},{"issuer_ca_id":295813,"issuer_name":"C=US, O=Let's Encrypt, CN=E7","common_name":"zap.mikoshi.de","name_value":"zap.mikoshi.de","id":20539914846,"entry_timestamp":"2025-08-24T10:34:58.16","not_before":"2025-08-24T09:36:27","not_after":"2025-11-22T09:36:26","serial_number":"062deaf87750f6cf54f0027bbf6f9617447b","result_count":2}]}
|
|
||||||
125
config.py
125
config.py
@@ -5,97 +5,110 @@ Handles API key storage, rate limiting, and default settings.
|
|||||||
|
|
||||||
import os
|
import os
|
||||||
from typing import Dict, Optional
|
from typing import Dict, Optional
|
||||||
from dotenv import load_dotenv
|
|
||||||
|
|
||||||
# Load environment variables from .env file
|
|
||||||
load_dotenv()
|
|
||||||
|
|
||||||
class Config:
|
class Config:
|
||||||
"""Configuration manager for DNSRecon application."""
|
"""Configuration manager for DNSRecon application."""
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
"""Initialize configuration with default values."""
|
"""Initialize configuration with default values."""
|
||||||
self.api_keys: Dict[str, Optional[str]] = {}
|
self.api_keys: Dict[str, Optional[str]] = {
|
||||||
|
'shodan': None
|
||||||
|
}
|
||||||
|
|
||||||
# --- General Settings ---
|
# Default settings
|
||||||
self.default_recursion_depth = 2
|
self.default_recursion_depth = 2
|
||||||
self.default_timeout = 30
|
self.default_timeout = 10
|
||||||
self.max_concurrent_requests = 5
|
self.max_concurrent_requests = 5
|
||||||
self.large_entity_threshold = 100
|
self.large_entity_threshold = 100
|
||||||
self.max_retries_per_target = 8
|
|
||||||
self.cache_expiry_hours = 12
|
|
||||||
|
|
||||||
# --- Provider Caching Settings ---
|
# Rate limiting settings (requests per minute)
|
||||||
self.cache_timeout_hours = 6 # Provider-specific cache timeout
|
|
||||||
|
|
||||||
# --- Rate Limiting (requests per minute) ---
|
|
||||||
self.rate_limits = {
|
self.rate_limits = {
|
||||||
'crtsh': 30,
|
'crtsh': 60, # Free service, be respectful
|
||||||
'shodan': 60,
|
'shodan': 60, # API dependent
|
||||||
'dns': 100
|
'dns': 100 # Local DNS queries
|
||||||
}
|
}
|
||||||
|
|
||||||
# --- Provider Settings ---
|
# Provider settings
|
||||||
self.enabled_providers = {
|
self.enabled_providers = {
|
||||||
'crtsh': True,
|
'crtsh': True, # Always enabled (free)
|
||||||
'dns': True,
|
'dns': True, # Always enabled (free)
|
||||||
'shodan': False
|
'shodan': False # Requires API key
|
||||||
}
|
}
|
||||||
|
|
||||||
# --- Logging ---
|
# Logging configuration
|
||||||
self.log_level = 'INFO'
|
self.log_level = 'INFO'
|
||||||
self.log_format = '%(asctime)s - %(name)s - %(levelname)s - %(message)s'
|
self.log_format = '%(asctime)s - %(name)s - %(levelname)s - %(message)s'
|
||||||
|
|
||||||
# --- Flask & Session Settings ---
|
# Flask configuration
|
||||||
self.flask_host = '127.0.0.1'
|
self.flask_host = '127.0.0.1'
|
||||||
self.flask_port = 5000
|
self.flask_port = 5000
|
||||||
self.flask_debug = True
|
self.flask_debug = True
|
||||||
self.flask_secret_key = 'default-secret-key-change-me'
|
|
||||||
self.flask_permanent_session_lifetime_hours = 2
|
|
||||||
self.session_timeout_minutes = 60
|
|
||||||
|
|
||||||
# Load environment variables to override defaults
|
def set_api_key(self, provider: str, api_key: str) -> bool:
|
||||||
self.load_from_env()
|
"""
|
||||||
|
Set API key for a provider.
|
||||||
|
|
||||||
def load_from_env(self):
|
Args:
|
||||||
"""Load configuration from environment variables."""
|
provider: Provider name (shodan, etc)
|
||||||
self.set_api_key('shodan', os.getenv('SHODAN_API_KEY'))
|
api_key: API key string
|
||||||
|
|
||||||
# Override settings from environment
|
Returns:
|
||||||
self.default_recursion_depth = int(os.getenv('DEFAULT_RECURSION_DEPTH', self.default_recursion_depth))
|
bool: True if key was set successfully
|
||||||
self.default_timeout = int(os.getenv('DEFAULT_TIMEOUT', self.default_timeout))
|
"""
|
||||||
self.max_concurrent_requests = int(os.getenv('MAX_CONCURRENT_REQUESTS', self.max_concurrent_requests))
|
if provider in self.api_keys:
|
||||||
self.large_entity_threshold = int(os.getenv('LARGE_ENTITY_THRESHOLD', self.large_entity_threshold))
|
self.api_keys[provider] = api_key
|
||||||
self.max_retries_per_target = int(os.getenv('MAX_RETRIES_PER_TARGET', self.max_retries_per_target))
|
self.enabled_providers[provider] = True if api_key else False
|
||||||
self.cache_expiry_hours = int(os.getenv('CACHE_EXPIRY_HOURS', self.cache_expiry_hours))
|
return True
|
||||||
self.cache_timeout_hours = int(os.getenv('CACHE_TIMEOUT_HOURS', self.cache_timeout_hours))
|
return False
|
||||||
|
|
||||||
# Override Flask and session settings
|
|
||||||
self.flask_host = os.getenv('FLASK_HOST', self.flask_host)
|
|
||||||
self.flask_port = int(os.getenv('FLASK_PORT', self.flask_port))
|
|
||||||
self.flask_debug = os.getenv('FLASK_DEBUG', str(self.flask_debug)).lower() == 'true'
|
|
||||||
self.flask_secret_key = os.getenv('FLASK_SECRET_KEY', self.flask_secret_key)
|
|
||||||
self.flask_permanent_session_lifetime_hours = int(os.getenv('FLASK_PERMANENT_SESSION_LIFETIME_HOURS', self.flask_permanent_session_lifetime_hours))
|
|
||||||
self.session_timeout_minutes = int(os.getenv('SESSION_TIMEOUT_MINUTES', self.session_timeout_minutes))
|
|
||||||
|
|
||||||
def set_api_key(self, provider: str, api_key: Optional[str]) -> bool:
|
|
||||||
"""Set API key for a provider."""
|
|
||||||
self.api_keys[provider] = api_key
|
|
||||||
if api_key:
|
|
||||||
self.enabled_providers[provider] = True
|
|
||||||
return True
|
|
||||||
|
|
||||||
def get_api_key(self, provider: str) -> Optional[str]:
|
def get_api_key(self, provider: str) -> Optional[str]:
|
||||||
"""Get API key for a provider."""
|
"""
|
||||||
|
Get API key for a provider.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider: Provider name
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
API key or None if not set
|
||||||
|
"""
|
||||||
return self.api_keys.get(provider)
|
return self.api_keys.get(provider)
|
||||||
|
|
||||||
def is_provider_enabled(self, provider: str) -> bool:
|
def is_provider_enabled(self, provider: str) -> bool:
|
||||||
"""Check if a provider is enabled."""
|
"""
|
||||||
|
Check if a provider is enabled.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider: Provider name
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
bool: True if provider is enabled
|
||||||
|
"""
|
||||||
return self.enabled_providers.get(provider, False)
|
return self.enabled_providers.get(provider, False)
|
||||||
|
|
||||||
def get_rate_limit(self, provider: str) -> int:
|
def get_rate_limit(self, provider: str) -> int:
|
||||||
"""Get rate limit for a provider."""
|
"""
|
||||||
|
Get rate limit for a provider.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider: Provider name
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Rate limit in requests per minute
|
||||||
|
"""
|
||||||
return self.rate_limits.get(provider, 60)
|
return self.rate_limits.get(provider, 60)
|
||||||
|
|
||||||
|
def load_from_env(self):
|
||||||
|
"""Load configuration from environment variables."""
|
||||||
|
if os.getenv('SHODAN_API_KEY'):
|
||||||
|
self.set_api_key('shodan', os.getenv('SHODAN_API_KEY'))
|
||||||
|
|
||||||
|
# Override default settings from environment
|
||||||
|
self.default_recursion_depth = int(os.getenv('DEFAULT_RECURSION_DEPTH', '2'))
|
||||||
|
self.flask_debug = os.getenv('FLASK_DEBUG', 'True').lower() == 'true'
|
||||||
|
self.default_timeout = 30
|
||||||
|
self.max_concurrent_requests = 5
|
||||||
|
|
||||||
|
|
||||||
# Global configuration instance
|
# Global configuration instance
|
||||||
config = Config()
|
config = Config()
|
||||||
@@ -8,6 +8,7 @@ from .scanner import Scanner, ScanStatus
|
|||||||
from .logger import ForensicLogger, get_forensic_logger, new_session
|
from .logger import ForensicLogger, get_forensic_logger, new_session
|
||||||
from .session_manager import session_manager
|
from .session_manager import session_manager
|
||||||
from .session_config import SessionConfig, create_session_config
|
from .session_config import SessionConfig, create_session_config
|
||||||
|
from .task_manager import TaskManager, TaskType, ReconTask
|
||||||
|
|
||||||
__all__ = [
|
__all__ = [
|
||||||
'GraphManager',
|
'GraphManager',
|
||||||
@@ -19,7 +20,10 @@ __all__ = [
|
|||||||
'new_session',
|
'new_session',
|
||||||
'session_manager',
|
'session_manager',
|
||||||
'SessionConfig',
|
'SessionConfig',
|
||||||
'create_session_config'
|
'create_session_config',
|
||||||
|
'TaskManager',
|
||||||
|
'TaskType',
|
||||||
|
'ReconTask'
|
||||||
]
|
]
|
||||||
|
|
||||||
__version__ = "1.0.0-phase2"
|
__version__ = "1.0.0-phase2"
|
||||||
@@ -1,5 +1,3 @@
|
|||||||
# core/graph_manager.py
|
|
||||||
|
|
||||||
"""
|
"""
|
||||||
Graph data model for DNSRecon using NetworkX.
|
Graph data model for DNSRecon using NetworkX.
|
||||||
Manages in-memory graph storage with confidence scoring and forensic metadata.
|
Manages in-memory graph storage with confidence scoring and forensic metadata.
|
||||||
@@ -52,23 +50,21 @@ class GraphManager:
|
|||||||
self.__dict__.update(state)
|
self.__dict__.update(state)
|
||||||
self.date_pattern = re.compile(r'^\d{4}-\d{2}-\d{2}[ T]\d{2}:\d{2}:\d{2}')
|
self.date_pattern = re.compile(r'^\d{4}-\d{2}-\d{2}[ T]\d{2}:\d{2}:\d{2}')
|
||||||
|
|
||||||
def _update_correlation_index(self, node_id: str, data: Any, path: List[str] = [], parent_attr: str = ""):
|
def _update_correlation_index(self, node_id: str, data: Any, path: List[str] = None):
|
||||||
"""Recursively traverse metadata and add hashable values to the index with better path tracking."""
|
"""Recursively traverse metadata and add hashable values to the index."""
|
||||||
if path is None:
|
if path is None:
|
||||||
path = []
|
path = []
|
||||||
|
|
||||||
if isinstance(data, dict):
|
if isinstance(data, dict):
|
||||||
for key, value in data.items():
|
for key, value in data.items():
|
||||||
self._update_correlation_index(node_id, value, path + [key], key)
|
self._update_correlation_index(node_id, value, path + [key])
|
||||||
elif isinstance(data, list):
|
elif isinstance(data, list):
|
||||||
for i, item in enumerate(data):
|
for i, item in enumerate(data):
|
||||||
# Instead of just using [i], include the parent attribute context
|
self._update_correlation_index(node_id, item, path + [f"[{i}]"])
|
||||||
list_path_component = f"[{i}]" if not parent_attr else f"{parent_attr}[{i}]"
|
|
||||||
self._update_correlation_index(node_id, item, path + [list_path_component], parent_attr)
|
|
||||||
else:
|
else:
|
||||||
self._add_to_correlation_index(node_id, data, ".".join(path), parent_attr)
|
self._add_to_correlation_index(node_id, data, ".".join(path))
|
||||||
|
|
||||||
def _add_to_correlation_index(self, node_id: str, value: Any, path_str: str, parent_attr: str = ""):
|
def _add_to_correlation_index(self, node_id: str, value: Any, path_str: str):
|
||||||
"""Add a hashable value to the correlation index, filtering out noise."""
|
"""Add a hashable value to the correlation index, filtering out noise."""
|
||||||
if not isinstance(value, (str, int, float, bool)) or value is None:
|
if not isinstance(value, (str, int, float, bool)) or value is None:
|
||||||
return
|
return
|
||||||
@@ -84,8 +80,8 @@ class GraphManager:
|
|||||||
return
|
return
|
||||||
if len(value) < 4 or value.lower() in ['true', 'false', 'unknown', 'none', 'crt.sh']:
|
if len(value) < 4 or value.lower() in ['true', 'false', 'unknown', 'none', 'crt.sh']:
|
||||||
return
|
return
|
||||||
elif isinstance(value, int) and (abs(value) < 1024 or abs(value) > 65535):
|
elif isinstance(value, int) and abs(value) < 9999:
|
||||||
return # Ignore small integers and common port numbers
|
return # Ignore small integers
|
||||||
elif isinstance(value, bool):
|
elif isinstance(value, bool):
|
||||||
return # Ignore boolean values
|
return # Ignore boolean values
|
||||||
|
|
||||||
@@ -94,47 +90,10 @@ class GraphManager:
|
|||||||
self.correlation_index[value] = {}
|
self.correlation_index[value] = {}
|
||||||
if node_id not in self.correlation_index[value]:
|
if node_id not in self.correlation_index[value]:
|
||||||
self.correlation_index[value][node_id] = []
|
self.correlation_index[value][node_id] = []
|
||||||
|
if path_str not in self.correlation_index[value][node_id]:
|
||||||
# Store both the full path and the parent attribute for better edge labeling
|
self.correlation_index[value][node_id].append(path_str)
|
||||||
correlation_entry = {
|
|
||||||
'path': path_str,
|
|
||||||
'parent_attr': parent_attr,
|
|
||||||
'meaningful_attr': self._extract_meaningful_attribute(path_str, parent_attr)
|
|
||||||
}
|
|
||||||
|
|
||||||
if correlation_entry not in self.correlation_index[value][node_id]:
|
|
||||||
self.correlation_index[value][node_id].append(correlation_entry)
|
|
||||||
|
|
||||||
def _extract_meaningful_attribute(self, path_str: str, parent_attr: str = "") -> str:
|
def _check_for_correlations(self, new_node_id: str, data: Any, path: List[str] = None) -> List[Dict]:
|
||||||
"""Extract the most meaningful attribute name from a path string."""
|
|
||||||
if not path_str:
|
|
||||||
return "unknown"
|
|
||||||
|
|
||||||
path_parts = path_str.split('.')
|
|
||||||
|
|
||||||
# Look for the last non-array-index part
|
|
||||||
for part in reversed(path_parts):
|
|
||||||
# Skip array indices like [0], [1], etc.
|
|
||||||
if not (part.startswith('[') and part.endswith(']') and part[1:-1].isdigit()):
|
|
||||||
# Clean up compound names like "hostnames[0]" to just "hostnames"
|
|
||||||
clean_part = re.sub(r'\[\d+\]$', '', part)
|
|
||||||
if clean_part:
|
|
||||||
return clean_part
|
|
||||||
|
|
||||||
# Fallback to parent attribute if available
|
|
||||||
if parent_attr:
|
|
||||||
return parent_attr
|
|
||||||
|
|
||||||
# Last resort - use the first meaningful part
|
|
||||||
for part in path_parts:
|
|
||||||
if not (part.startswith('[') and part.endswith(']') and part[1:-1].isdigit()):
|
|
||||||
clean_part = re.sub(r'\[\d+\]$', '', part)
|
|
||||||
if clean_part:
|
|
||||||
return clean_part
|
|
||||||
|
|
||||||
return "correlation"
|
|
||||||
|
|
||||||
def _check_for_correlations(self, new_node_id: str, data: Any, path: List[str] = [], parent_attr: str = "") -> List[Dict]:
|
|
||||||
"""Recursively traverse metadata to find correlations with existing data."""
|
"""Recursively traverse metadata to find correlations with existing data."""
|
||||||
if path is None:
|
if path is None:
|
||||||
path = []
|
path = []
|
||||||
@@ -144,11 +103,10 @@ class GraphManager:
|
|||||||
for key, value in data.items():
|
for key, value in data.items():
|
||||||
if key == 'source': # Avoid correlating on the provider name
|
if key == 'source': # Avoid correlating on the provider name
|
||||||
continue
|
continue
|
||||||
all_correlations.extend(self._check_for_correlations(new_node_id, value, path + [key], key))
|
all_correlations.extend(self._check_for_correlations(new_node_id, value, path + [key]))
|
||||||
elif isinstance(data, list):
|
elif isinstance(data, list):
|
||||||
for i, item in enumerate(data):
|
for i, item in enumerate(data):
|
||||||
list_path_component = f"[{i}]" if not parent_attr else f"{parent_attr}[{i}]"
|
all_correlations.extend(self._check_for_correlations(new_node_id, item, path + [f"[{i}]"]))
|
||||||
all_correlations.extend(self._check_for_correlations(new_node_id, item, path + [list_path_component], parent_attr))
|
|
||||||
else:
|
else:
|
||||||
value = data
|
value = data
|
||||||
if value in self.correlation_index:
|
if value in self.correlation_index:
|
||||||
@@ -159,31 +117,11 @@ class GraphManager:
|
|||||||
if len(unique_nodes) < 2:
|
if len(unique_nodes) < 2:
|
||||||
return all_correlations # Correlation must involve at least two distinct nodes
|
return all_correlations # Correlation must involve at least two distinct nodes
|
||||||
|
|
||||||
new_source = {
|
new_source = {'node_id': new_node_id, 'path': ".".join(path)}
|
||||||
'node_id': new_node_id,
|
|
||||||
'path': ".".join(path),
|
|
||||||
'parent_attr': parent_attr,
|
|
||||||
'meaningful_attr': self._extract_meaningful_attribute(".".join(path), parent_attr)
|
|
||||||
}
|
|
||||||
all_sources = [new_source]
|
all_sources = [new_source]
|
||||||
|
for node_id, paths in existing_nodes_with_paths.items():
|
||||||
for node_id, path_entries in existing_nodes_with_paths.items():
|
for p_str in paths:
|
||||||
for entry in path_entries:
|
all_sources.append({'node_id': node_id, 'path': p_str})
|
||||||
if isinstance(entry, dict):
|
|
||||||
all_sources.append({
|
|
||||||
'node_id': node_id,
|
|
||||||
'path': entry['path'],
|
|
||||||
'parent_attr': entry.get('parent_attr', ''),
|
|
||||||
'meaningful_attr': entry.get('meaningful_attr', self._extract_meaningful_attribute(entry['path'], entry.get('parent_attr', '')))
|
|
||||||
})
|
|
||||||
else:
|
|
||||||
# Handle legacy string-only entries
|
|
||||||
all_sources.append({
|
|
||||||
'node_id': node_id,
|
|
||||||
'path': str(entry),
|
|
||||||
'parent_attr': '',
|
|
||||||
'meaningful_attr': self._extract_meaningful_attribute(str(entry))
|
|
||||||
})
|
|
||||||
|
|
||||||
all_correlations.append({
|
all_correlations.append({
|
||||||
'value': value,
|
'value': value,
|
||||||
@@ -225,7 +163,8 @@ class GraphManager:
|
|||||||
# Skip creating correlation node - would be redundant
|
# Skip creating correlation node - would be redundant
|
||||||
continue
|
continue
|
||||||
|
|
||||||
eligible_nodes = set(corr['nodes'])
|
# STEP 2: Filter out node pairs that already have direct edges
|
||||||
|
eligible_nodes = self._filter_nodes_without_direct_edges(set(corr['nodes']))
|
||||||
|
|
||||||
if len(eligible_nodes) < 2:
|
if len(eligible_nodes) < 2:
|
||||||
# Need at least 2 nodes to create a correlation
|
# Need at least 2 nodes to create a correlation
|
||||||
@@ -245,12 +184,11 @@ class GraphManager:
|
|||||||
metadata={'values': [value], 'sources': corr['sources'],
|
metadata={'values': [value], 'sources': corr['sources'],
|
||||||
'correlated_nodes': list(eligible_nodes)})
|
'correlated_nodes': list(eligible_nodes)})
|
||||||
|
|
||||||
# Create edges from eligible nodes to this correlation node with better labeling
|
# Create edges from eligible nodes to this correlation node
|
||||||
for c_node_id in eligible_nodes:
|
for c_node_id in eligible_nodes:
|
||||||
if self.graph.has_node(c_node_id):
|
if self.graph.has_node(c_node_id):
|
||||||
# Find the best attribute name for this node
|
attribute = corr['sources'][0]['path'].split('.')[-1]
|
||||||
meaningful_attr = self._find_best_attribute_name_for_node(c_node_id, corr['sources'])
|
relationship_type = f"c_{attribute}"
|
||||||
relationship_type = f"c_{meaningful_attr}"
|
|
||||||
self.add_edge(c_node_id, correlation_node_id, relationship_type, confidence_score=0.9)
|
self.add_edge(c_node_id, correlation_node_id, relationship_type, confidence_score=0.9)
|
||||||
|
|
||||||
self._update_correlation_index(node_id, attributes)
|
self._update_correlation_index(node_id, attributes)
|
||||||
@@ -258,34 +196,27 @@ class GraphManager:
|
|||||||
self.last_modified = datetime.now(timezone.utc).isoformat()
|
self.last_modified = datetime.now(timezone.utc).isoformat()
|
||||||
return is_new_node
|
return is_new_node
|
||||||
|
|
||||||
def _find_best_attribute_name_for_node(self, node_id: str, sources: List[Dict]) -> str:
|
def _filter_nodes_without_direct_edges(self, node_set: set) -> set:
|
||||||
"""Find the best attribute name for a correlation edge by looking at the sources."""
|
"""
|
||||||
node_sources = [s for s in sources if s['node_id'] == node_id]
|
Filter out nodes that already have direct edges between them.
|
||||||
|
Returns set of nodes that should be included in correlation.
|
||||||
|
"""
|
||||||
|
nodes_list = list(node_set)
|
||||||
|
eligible_nodes = set(node_set) # Start with all nodes
|
||||||
|
|
||||||
if not node_sources:
|
# Check all pairs of nodes
|
||||||
return "correlation"
|
for i in range(len(nodes_list)):
|
||||||
|
for j in range(i + 1, len(nodes_list)):
|
||||||
|
node_a = nodes_list[i]
|
||||||
|
node_b = nodes_list[j]
|
||||||
|
|
||||||
|
# Check if direct edge exists in either direction
|
||||||
|
if self._has_direct_edge_bidirectional(node_a, node_b):
|
||||||
|
# Remove both nodes from eligible set since they're already connected
|
||||||
|
eligible_nodes.discard(node_a)
|
||||||
|
eligible_nodes.discard(node_b)
|
||||||
|
|
||||||
# Use the meaningful_attr if available
|
return eligible_nodes
|
||||||
for source in node_sources:
|
|
||||||
meaningful_attr = source.get('meaningful_attr')
|
|
||||||
if meaningful_attr and meaningful_attr != "unknown":
|
|
||||||
return meaningful_attr
|
|
||||||
|
|
||||||
# Fallback to parent_attr
|
|
||||||
for source in node_sources:
|
|
||||||
parent_attr = source.get('parent_attr')
|
|
||||||
if parent_attr:
|
|
||||||
return parent_attr
|
|
||||||
|
|
||||||
# Last resort - extract from path
|
|
||||||
for source in node_sources:
|
|
||||||
path = source.get('path', '')
|
|
||||||
if path:
|
|
||||||
extracted = self._extract_meaningful_attribute(path)
|
|
||||||
if extracted != "unknown":
|
|
||||||
return extracted
|
|
||||||
|
|
||||||
return "correlation"
|
|
||||||
|
|
||||||
def _has_direct_edge_bidirectional(self, node_a: str, node_b: str) -> bool:
|
def _has_direct_edge_bidirectional(self, node_a: str, node_b: str) -> bool:
|
||||||
"""
|
"""
|
||||||
@@ -359,7 +290,7 @@ class GraphManager:
|
|||||||
# Create set of unique sources based on (node_id, path) tuples
|
# Create set of unique sources based on (node_id, path) tuples
|
||||||
source_set = set()
|
source_set = set()
|
||||||
for source in existing_sources + new_sources:
|
for source in existing_sources + new_sources:
|
||||||
source_tuple = (source['node_id'], source.get('path', ''))
|
source_tuple = (source['node_id'], source['path'])
|
||||||
source_set.add(source_tuple)
|
source_set.add(source_tuple)
|
||||||
|
|
||||||
# Convert back to list of dictionaries
|
# Convert back to list of dictionaries
|
||||||
@@ -414,29 +345,6 @@ class GraphManager:
|
|||||||
self.last_modified = datetime.now(timezone.utc).isoformat()
|
self.last_modified = datetime.now(timezone.utc).isoformat()
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def remove_node(self, node_id: str) -> bool:
|
|
||||||
"""Remove a node and its connected edges from the graph."""
|
|
||||||
if not self.graph.has_node(node_id):
|
|
||||||
return False
|
|
||||||
|
|
||||||
# Remove node from the graph (NetworkX handles removing connected edges)
|
|
||||||
self.graph.remove_node(node_id)
|
|
||||||
|
|
||||||
# Clean up the correlation index
|
|
||||||
keys_to_delete = []
|
|
||||||
for value, nodes in self.correlation_index.items():
|
|
||||||
if node_id in nodes:
|
|
||||||
del nodes[node_id]
|
|
||||||
if not nodes: # If no other nodes are associated with this value, remove it
|
|
||||||
keys_to_delete.append(value)
|
|
||||||
|
|
||||||
for key in keys_to_delete:
|
|
||||||
if key in self.correlation_index:
|
|
||||||
del self.correlation_index[key]
|
|
||||||
|
|
||||||
self.last_modified = datetime.now(timezone.utc).isoformat()
|
|
||||||
return True
|
|
||||||
|
|
||||||
def get_node_count(self) -> int:
|
def get_node_count(self) -> int:
|
||||||
"""Get total number of nodes in the graph."""
|
"""Get total number of nodes in the graph."""
|
||||||
return self.graph.number_of_nodes()
|
return self.graph.number_of_nodes()
|
||||||
@@ -513,14 +421,10 @@ class GraphManager:
|
|||||||
def _get_confidence_distribution(self) -> Dict[str, int]:
|
def _get_confidence_distribution(self) -> Dict[str, int]:
|
||||||
"""Get distribution of edge confidence scores."""
|
"""Get distribution of edge confidence scores."""
|
||||||
distribution = {'high': 0, 'medium': 0, 'low': 0}
|
distribution = {'high': 0, 'medium': 0, 'low': 0}
|
||||||
for _, _, data in self.graph.edges(data=True):
|
for _, _, confidence in self.graph.edges(data='confidence_score', default=0):
|
||||||
confidence = data.get('confidence_score', 0)
|
if confidence >= 0.8: distribution['high'] += 1
|
||||||
if confidence >= 0.8:
|
elif confidence >= 0.6: distribution['medium'] += 1
|
||||||
distribution['high'] += 1
|
else: distribution['low'] += 1
|
||||||
elif confidence >= 0.6:
|
|
||||||
distribution['medium'] += 1
|
|
||||||
else:
|
|
||||||
distribution['low'] += 1
|
|
||||||
return distribution
|
return distribution
|
||||||
|
|
||||||
def get_statistics(self) -> Dict[str, Any]:
|
def get_statistics(self) -> Dict[str, Any]:
|
||||||
@@ -535,10 +439,9 @@ class GraphManager:
|
|||||||
# Calculate distributions
|
# Calculate distributions
|
||||||
for node_type in NodeType:
|
for node_type in NodeType:
|
||||||
stats['node_type_distribution'][node_type.value] = self.get_nodes_by_type(node_type).__len__()
|
stats['node_type_distribution'][node_type.value] = self.get_nodes_by_type(node_type).__len__()
|
||||||
for _, _, data in self.graph.edges(data=True):
|
for _, _, rel_type in self.graph.edges(data='relationship_type', default='unknown'):
|
||||||
rel_type = data.get('relationship_type', 'unknown')
|
|
||||||
stats['relationship_type_distribution'][rel_type] = stats['relationship_type_distribution'].get(rel_type, 0) + 1
|
stats['relationship_type_distribution'][rel_type] = stats['relationship_type_distribution'].get(rel_type, 0) + 1
|
||||||
provider = data.get('source_provider', 'unknown')
|
for _, _, provider in self.graph.edges(data='source_provider', default='unknown'):
|
||||||
stats['provider_distribution'][provider] = stats['provider_distribution'].get(provider, 0) + 1
|
stats['provider_distribution'][provider] = stats['provider_distribution'].get(provider, 0) + 1
|
||||||
return stats
|
return stats
|
||||||
|
|
||||||
|
|||||||
@@ -42,7 +42,7 @@ class ForensicLogger:
|
|||||||
Maintains detailed audit trail of all reconnaissance activities.
|
Maintains detailed audit trail of all reconnaissance activities.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, session_id: str = ""):
|
def __init__(self, session_id: str = None):
|
||||||
"""
|
"""
|
||||||
Initialize forensic logger.
|
Initialize forensic logger.
|
||||||
|
|
||||||
@@ -203,6 +203,8 @@ class ForensicLogger:
|
|||||||
self.session_metadata['target_domains'] = list(self.session_metadata['target_domains'])
|
self.session_metadata['target_domains'] = list(self.session_metadata['target_domains'])
|
||||||
|
|
||||||
self.logger.info(f"Scan Complete - Session: {self.session_id}")
|
self.logger.info(f"Scan Complete - Session: {self.session_id}")
|
||||||
|
self.logger.info(f"Total API Requests: {self.session_metadata['total_requests']}")
|
||||||
|
self.logger.info(f"Total Relationships: {self.session_metadata['total_relationships']}")
|
||||||
|
|
||||||
def export_audit_trail(self) -> Dict[str, Any]:
|
def export_audit_trail(self) -> Dict[str, Any]:
|
||||||
"""
|
"""
|
||||||
|
|||||||
1120
core/scanner.py
1120
core/scanner.py
File diff suppressed because it is too large
Load Diff
@@ -1,20 +1,372 @@
|
|||||||
"""
|
"""
|
||||||
Per-session configuration management for DNSRecon.
|
Enhanced per-session configuration management for DNSRecon.
|
||||||
Provides isolated configuration instances for each user session.
|
Provides isolated configuration instances for each user session while supporting global caching.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
from config import Config
|
import os
|
||||||
|
from typing import Dict, Optional
|
||||||
|
|
||||||
class SessionConfig(Config):
|
|
||||||
|
class SessionConfig:
|
||||||
"""
|
"""
|
||||||
Session-specific configuration that inherits from global config
|
Enhanced session-specific configuration that inherits from global config
|
||||||
but maintains isolated API keys and provider settings.
|
but maintains isolated API keys and provider settings while supporting global caching.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
"""Initialize session config with global defaults."""
|
"""Initialize enhanced session config with global cache support."""
|
||||||
super().__init__()
|
# Copy all attributes from global config
|
||||||
|
self.api_keys: Dict[str, Optional[str]] = {
|
||||||
|
'shodan': None
|
||||||
|
}
|
||||||
|
|
||||||
|
# Default settings (copied from global config)
|
||||||
|
self.default_recursion_depth = 2
|
||||||
|
self.default_timeout = 30
|
||||||
|
self.max_concurrent_requests = 5
|
||||||
|
self.large_entity_threshold = 100
|
||||||
|
|
||||||
|
# Enhanced rate limiting settings (per session)
|
||||||
|
self.rate_limits = {
|
||||||
|
'crtsh': 60,
|
||||||
|
'shodan': 60,
|
||||||
|
'dns': 100
|
||||||
|
}
|
||||||
|
|
||||||
|
# Enhanced provider settings (per session)
|
||||||
|
self.enabled_providers = {
|
||||||
|
'crtsh': True,
|
||||||
|
'dns': True,
|
||||||
|
'shodan': False
|
||||||
|
}
|
||||||
|
|
||||||
|
# Task-based execution settings
|
||||||
|
self.task_retry_settings = {
|
||||||
|
'max_retries': 3,
|
||||||
|
'base_backoff_seconds': 1.0,
|
||||||
|
'max_backoff_seconds': 60.0,
|
||||||
|
'retry_on_rate_limit': True,
|
||||||
|
'retry_on_connection_error': True,
|
||||||
|
'retry_on_timeout': True
|
||||||
|
}
|
||||||
|
|
||||||
|
# Cache settings (global across all sessions)
|
||||||
|
self.cache_settings = {
|
||||||
|
'enabled': True,
|
||||||
|
'expiry_hours': 12,
|
||||||
|
'cache_base_dir': '.cache',
|
||||||
|
'per_provider_directories': True,
|
||||||
|
'thread_safe_operations': True
|
||||||
|
}
|
||||||
|
|
||||||
|
# Logging configuration
|
||||||
|
self.log_level = 'INFO'
|
||||||
|
self.log_format = '%(asctime)s - %(name)s - %(levelname)s - %(message)s'
|
||||||
|
|
||||||
|
# Flask configuration (shared)
|
||||||
|
self.flask_host = '127.0.0.1'
|
||||||
|
self.flask_port = 5000
|
||||||
|
self.flask_debug = True
|
||||||
|
|
||||||
|
# Session isolation settings
|
||||||
|
self.session_isolation = {
|
||||||
|
'enforce_single_session_per_user': True,
|
||||||
|
'consolidate_session_data_on_replacement': True,
|
||||||
|
'user_fingerprinting_enabled': True,
|
||||||
|
'session_timeout_minutes': 60
|
||||||
|
}
|
||||||
|
|
||||||
|
# Circuit breaker settings for provider reliability
|
||||||
|
self.circuit_breaker = {
|
||||||
|
'enabled': True,
|
||||||
|
'failure_threshold': 5, # Failures before opening circuit
|
||||||
|
'recovery_timeout_seconds': 300, # 5 minutes before trying again
|
||||||
|
'half_open_max_calls': 3 # Test calls when recovering
|
||||||
|
}
|
||||||
|
|
||||||
|
def set_api_key(self, provider: str, api_key: str) -> bool:
|
||||||
|
"""
|
||||||
|
Set API key for a provider in this session.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider: Provider name (shodan, etc)
|
||||||
|
api_key: API key string (empty string to clear)
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
bool: True if key was set successfully
|
||||||
|
"""
|
||||||
|
if provider in self.api_keys:
|
||||||
|
# Handle clearing of API keys
|
||||||
|
if api_key and api_key.strip():
|
||||||
|
self.api_keys[provider] = api_key.strip()
|
||||||
|
self.enabled_providers[provider] = True
|
||||||
|
else:
|
||||||
|
self.api_keys[provider] = None
|
||||||
|
self.enabled_providers[provider] = False
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
def get_api_key(self, provider: str) -> Optional[str]:
|
||||||
|
"""
|
||||||
|
Get API key for a provider in this session.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider: Provider name
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
API key or None if not set
|
||||||
|
"""
|
||||||
|
return self.api_keys.get(provider)
|
||||||
|
|
||||||
|
def is_provider_enabled(self, provider: str) -> bool:
|
||||||
|
"""
|
||||||
|
Check if a provider is enabled in this session.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider: Provider name
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
bool: True if provider is enabled
|
||||||
|
"""
|
||||||
|
return self.enabled_providers.get(provider, False)
|
||||||
|
|
||||||
|
def get_rate_limit(self, provider: str) -> int:
|
||||||
|
"""
|
||||||
|
Get rate limit for a provider in this session.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider: Provider name
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Rate limit in requests per minute
|
||||||
|
"""
|
||||||
|
return self.rate_limits.get(provider, 60)
|
||||||
|
|
||||||
|
def get_task_retry_config(self) -> Dict[str, any]:
|
||||||
|
"""
|
||||||
|
Get task retry configuration for this session.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Dictionary with retry settings
|
||||||
|
"""
|
||||||
|
return self.task_retry_settings.copy()
|
||||||
|
|
||||||
|
def get_cache_config(self) -> Dict[str, any]:
|
||||||
|
"""
|
||||||
|
Get cache configuration (global settings).
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Dictionary with cache settings
|
||||||
|
"""
|
||||||
|
return self.cache_settings.copy()
|
||||||
|
|
||||||
|
def is_circuit_breaker_enabled(self) -> bool:
|
||||||
|
"""Check if circuit breaker is enabled for provider reliability."""
|
||||||
|
return self.circuit_breaker.get('enabled', True)
|
||||||
|
|
||||||
|
def get_circuit_breaker_config(self) -> Dict[str, any]:
|
||||||
|
"""Get circuit breaker configuration."""
|
||||||
|
return self.circuit_breaker.copy()
|
||||||
|
|
||||||
|
def update_provider_settings(self, provider_updates: Dict[str, Dict[str, any]]) -> bool:
|
||||||
|
"""
|
||||||
|
Update provider-specific settings in bulk.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider_updates: Dictionary of provider -> settings updates
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
bool: True if updates were applied successfully
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
for provider_name, updates in provider_updates.items():
|
||||||
|
# Update rate limits
|
||||||
|
if 'rate_limit' in updates:
|
||||||
|
self.rate_limits[provider_name] = updates['rate_limit']
|
||||||
|
|
||||||
|
# Update enabled status
|
||||||
|
if 'enabled' in updates:
|
||||||
|
self.enabled_providers[provider_name] = updates['enabled']
|
||||||
|
|
||||||
|
# Update API key
|
||||||
|
if 'api_key' in updates:
|
||||||
|
self.set_api_key(provider_name, updates['api_key'])
|
||||||
|
|
||||||
|
return True
|
||||||
|
except Exception as e:
|
||||||
|
print(f"Error updating provider settings: {e}")
|
||||||
|
return False
|
||||||
|
|
||||||
|
def validate_configuration(self) -> Dict[str, any]:
|
||||||
|
"""
|
||||||
|
Validate the current configuration and return validation results.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Dictionary with validation results and any issues found
|
||||||
|
"""
|
||||||
|
validation_result = {
|
||||||
|
'valid': True,
|
||||||
|
'warnings': [],
|
||||||
|
'errors': [],
|
||||||
|
'provider_status': {}
|
||||||
|
}
|
||||||
|
|
||||||
|
# Validate provider configurations
|
||||||
|
for provider_name, enabled in self.enabled_providers.items():
|
||||||
|
provider_status = {
|
||||||
|
'enabled': enabled,
|
||||||
|
'has_api_key': bool(self.api_keys.get(provider_name)),
|
||||||
|
'rate_limit': self.rate_limits.get(provider_name, 60)
|
||||||
|
}
|
||||||
|
|
||||||
|
# Check for potential issues
|
||||||
|
if enabled and provider_name in ['shodan'] and not provider_status['has_api_key']:
|
||||||
|
validation_result['warnings'].append(
|
||||||
|
f"Provider '{provider_name}' is enabled but missing API key"
|
||||||
|
)
|
||||||
|
|
||||||
|
validation_result['provider_status'][provider_name] = provider_status
|
||||||
|
|
||||||
|
# Validate task settings
|
||||||
|
if self.task_retry_settings['max_retries'] > 10:
|
||||||
|
validation_result['warnings'].append(
|
||||||
|
f"High retry count ({self.task_retry_settings['max_retries']}) may cause long delays"
|
||||||
|
)
|
||||||
|
|
||||||
|
# Validate concurrent settings
|
||||||
|
if self.max_concurrent_requests > 10:
|
||||||
|
validation_result['warnings'].append(
|
||||||
|
f"High concurrency ({self.max_concurrent_requests}) may overwhelm providers"
|
||||||
|
)
|
||||||
|
|
||||||
|
# Validate cache settings
|
||||||
|
if not os.path.exists(self.cache_settings['cache_base_dir']):
|
||||||
|
try:
|
||||||
|
os.makedirs(self.cache_settings['cache_base_dir'], exist_ok=True)
|
||||||
|
except Exception as e:
|
||||||
|
validation_result['errors'].append(f"Cannot create cache directory: {e}")
|
||||||
|
validation_result['valid'] = False
|
||||||
|
|
||||||
|
return validation_result
|
||||||
|
|
||||||
|
def load_from_env(self):
|
||||||
|
"""Load configuration from environment variables with enhanced validation."""
|
||||||
|
# Load API keys from environment
|
||||||
|
if os.getenv('SHODAN_API_KEY') and not self.api_keys['shodan']:
|
||||||
|
self.set_api_key('shodan', os.getenv('SHODAN_API_KEY'))
|
||||||
|
print("Loaded Shodan API key from environment")
|
||||||
|
|
||||||
|
# Override default settings from environment
|
||||||
|
self.default_recursion_depth = int(os.getenv('DEFAULT_RECURSION_DEPTH', '2'))
|
||||||
|
self.default_timeout = int(os.getenv('DEFAULT_TIMEOUT', '30'))
|
||||||
|
self.max_concurrent_requests = int(os.getenv('MAX_CONCURRENT_REQUESTS', '5'))
|
||||||
|
|
||||||
|
# Load task retry settings from environment
|
||||||
|
if os.getenv('TASK_MAX_RETRIES'):
|
||||||
|
self.task_retry_settings['max_retries'] = int(os.getenv('TASK_MAX_RETRIES'))
|
||||||
|
|
||||||
|
if os.getenv('TASK_BASE_BACKOFF'):
|
||||||
|
self.task_retry_settings['base_backoff_seconds'] = float(os.getenv('TASK_BASE_BACKOFF'))
|
||||||
|
|
||||||
|
# Load cache settings from environment
|
||||||
|
if os.getenv('CACHE_EXPIRY_HOURS'):
|
||||||
|
self.cache_settings['expiry_hours'] = int(os.getenv('CACHE_EXPIRY_HOURS'))
|
||||||
|
|
||||||
|
if os.getenv('CACHE_DISABLED'):
|
||||||
|
self.cache_settings['enabled'] = os.getenv('CACHE_DISABLED').lower() != 'true'
|
||||||
|
|
||||||
|
# Load circuit breaker settings
|
||||||
|
if os.getenv('CIRCUIT_BREAKER_DISABLED'):
|
||||||
|
self.circuit_breaker['enabled'] = os.getenv('CIRCUIT_BREAKER_DISABLED').lower() != 'true'
|
||||||
|
|
||||||
|
# Flask settings
|
||||||
|
self.flask_debug = os.getenv('FLASK_DEBUG', 'True').lower() == 'true'
|
||||||
|
|
||||||
|
print("Enhanced configuration loaded from environment")
|
||||||
|
|
||||||
|
def export_config_summary(self) -> Dict[str, any]:
|
||||||
|
"""
|
||||||
|
Export a summary of the current configuration for debugging/logging.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Dictionary with configuration summary (API keys redacted)
|
||||||
|
"""
|
||||||
|
return {
|
||||||
|
'providers': {
|
||||||
|
provider: {
|
||||||
|
'enabled': self.enabled_providers.get(provider, False),
|
||||||
|
'has_api_key': bool(self.api_keys.get(provider)),
|
||||||
|
'rate_limit': self.rate_limits.get(provider, 60)
|
||||||
|
}
|
||||||
|
for provider in self.enabled_providers.keys()
|
||||||
|
},
|
||||||
|
'task_settings': {
|
||||||
|
'max_retries': self.task_retry_settings['max_retries'],
|
||||||
|
'max_concurrent_requests': self.max_concurrent_requests,
|
||||||
|
'large_entity_threshold': self.large_entity_threshold
|
||||||
|
},
|
||||||
|
'cache_settings': {
|
||||||
|
'enabled': self.cache_settings['enabled'],
|
||||||
|
'expiry_hours': self.cache_settings['expiry_hours'],
|
||||||
|
'base_directory': self.cache_settings['cache_base_dir']
|
||||||
|
},
|
||||||
|
'session_settings': {
|
||||||
|
'isolation_enabled': self.session_isolation['enforce_single_session_per_user'],
|
||||||
|
'consolidation_enabled': self.session_isolation['consolidate_session_data_on_replacement'],
|
||||||
|
'timeout_minutes': self.session_isolation['session_timeout_minutes']
|
||||||
|
},
|
||||||
|
'circuit_breaker': {
|
||||||
|
'enabled': self.circuit_breaker['enabled'],
|
||||||
|
'failure_threshold': self.circuit_breaker['failure_threshold'],
|
||||||
|
'recovery_timeout': self.circuit_breaker['recovery_timeout_seconds']
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
def create_session_config() -> 'SessionConfig':
|
|
||||||
"""Create a new session configuration instance."""
|
def create_session_config() -> SessionConfig:
|
||||||
return SessionConfig()
|
"""
|
||||||
|
Create a new enhanced session configuration instance.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Configured SessionConfig instance
|
||||||
|
"""
|
||||||
|
session_config = SessionConfig()
|
||||||
|
session_config.load_from_env()
|
||||||
|
|
||||||
|
# Validate configuration and log any issues
|
||||||
|
validation = session_config.validate_configuration()
|
||||||
|
if validation['warnings']:
|
||||||
|
print("Configuration warnings:")
|
||||||
|
for warning in validation['warnings']:
|
||||||
|
print(f" WARNING: {warning}")
|
||||||
|
|
||||||
|
if validation['errors']:
|
||||||
|
print("Configuration errors:")
|
||||||
|
for error in validation['errors']:
|
||||||
|
print(f" ERROR: {error}")
|
||||||
|
|
||||||
|
if not validation['valid']:
|
||||||
|
raise ValueError("Configuration validation failed - see errors above")
|
||||||
|
|
||||||
|
print(f"Enhanced session configuration created successfully")
|
||||||
|
return session_config
|
||||||
|
|
||||||
|
|
||||||
|
def create_test_config() -> SessionConfig:
|
||||||
|
"""
|
||||||
|
Create a test configuration with safe defaults for testing.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Test-safe SessionConfig instance
|
||||||
|
"""
|
||||||
|
test_config = SessionConfig()
|
||||||
|
|
||||||
|
# Override settings for testing
|
||||||
|
test_config.max_concurrent_requests = 2
|
||||||
|
test_config.task_retry_settings['max_retries'] = 1
|
||||||
|
test_config.task_retry_settings['base_backoff_seconds'] = 0.1
|
||||||
|
test_config.cache_settings['expiry_hours'] = 1
|
||||||
|
test_config.session_isolation['session_timeout_minutes'] = 10
|
||||||
|
|
||||||
|
print("Test configuration created")
|
||||||
|
return test_config
|
||||||
@@ -5,41 +5,153 @@ import time
|
|||||||
import uuid
|
import uuid
|
||||||
import redis
|
import redis
|
||||||
import pickle
|
import pickle
|
||||||
from typing import Dict, Optional, Any, List
|
import hashlib
|
||||||
|
from typing import Dict, Optional, Any, List, Tuple
|
||||||
|
|
||||||
from core.scanner import Scanner
|
from core.scanner import Scanner
|
||||||
from config import config
|
|
||||||
|
|
||||||
# WARNING: Using pickle can be a security risk if the data source is not trusted.
|
|
||||||
# In this case, we are only serializing/deserializing our own trusted Scanner objects,
|
class UserIdentifier:
|
||||||
# which is generally safe. Do not unpickle data from untrusted sources.
|
"""Handles user identification for session management."""
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def generate_user_fingerprint(client_ip: str, user_agent: str) -> str:
|
||||||
|
"""
|
||||||
|
Generate a unique fingerprint for a user based on IP and User-Agent.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
client_ip: Client IP address
|
||||||
|
user_agent: User-Agent header value
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Unique user fingerprint hash
|
||||||
|
"""
|
||||||
|
# Create deterministic user identifier
|
||||||
|
user_data = f"{client_ip}:{user_agent[:100]}" # Limit UA to 100 chars
|
||||||
|
fingerprint = hashlib.sha256(user_data.encode()).hexdigest()[:16] # 16 char fingerprint
|
||||||
|
return f"user_{fingerprint}"
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def extract_request_info(request) -> Tuple[str, str]:
|
||||||
|
"""
|
||||||
|
Extract client IP and User-Agent from Flask request.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
request: Flask request object
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tuple of (client_ip, user_agent)
|
||||||
|
"""
|
||||||
|
# Handle proxy headers for real IP
|
||||||
|
client_ip = request.headers.get('X-Forwarded-For', '').split(',')[0].strip()
|
||||||
|
if not client_ip:
|
||||||
|
client_ip = request.headers.get('X-Real-IP', '')
|
||||||
|
if not client_ip:
|
||||||
|
client_ip = request.remote_addr or 'unknown'
|
||||||
|
|
||||||
|
user_agent = request.headers.get('User-Agent', 'unknown')
|
||||||
|
|
||||||
|
return client_ip, user_agent
|
||||||
|
|
||||||
|
|
||||||
|
class SessionConsolidator:
|
||||||
|
"""Handles consolidation of session data when replacing sessions."""
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def consolidate_scanner_data(old_scanner: 'Scanner', new_scanner: 'Scanner') -> 'Scanner':
|
||||||
|
"""
|
||||||
|
Consolidate useful data from old scanner into new scanner.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
old_scanner: Scanner from terminated session
|
||||||
|
new_scanner: New scanner instance
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Enhanced new scanner with consolidated data
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
# Consolidate graph data if old scanner has valuable data
|
||||||
|
if old_scanner and hasattr(old_scanner, 'graph') and old_scanner.graph:
|
||||||
|
old_stats = old_scanner.graph.get_statistics()
|
||||||
|
if old_stats['basic_metrics']['total_nodes'] > 0:
|
||||||
|
print(f"Consolidating graph data: {old_stats['basic_metrics']['total_nodes']} nodes, {old_stats['basic_metrics']['total_edges']} edges")
|
||||||
|
|
||||||
|
# Transfer nodes and edges to new scanner's graph
|
||||||
|
for node_id, node_data in old_scanner.graph.graph.nodes(data=True):
|
||||||
|
# Add node to new graph with all attributes
|
||||||
|
new_scanner.graph.graph.add_node(node_id, **node_data)
|
||||||
|
|
||||||
|
for source, target, edge_data in old_scanner.graph.graph.edges(data=True):
|
||||||
|
# Add edge to new graph with all attributes
|
||||||
|
new_scanner.graph.graph.add_edge(source, target, **edge_data)
|
||||||
|
|
||||||
|
# Update correlation index
|
||||||
|
if hasattr(old_scanner.graph, 'correlation_index'):
|
||||||
|
new_scanner.graph.correlation_index = old_scanner.graph.correlation_index.copy()
|
||||||
|
|
||||||
|
# Update timestamps
|
||||||
|
new_scanner.graph.creation_time = old_scanner.graph.creation_time
|
||||||
|
new_scanner.graph.last_modified = old_scanner.graph.last_modified
|
||||||
|
|
||||||
|
# Consolidate provider statistics
|
||||||
|
if old_scanner and hasattr(old_scanner, 'providers') and old_scanner.providers:
|
||||||
|
for old_provider in old_scanner.providers:
|
||||||
|
# Find matching provider in new scanner
|
||||||
|
matching_new_provider = None
|
||||||
|
for new_provider in new_scanner.providers:
|
||||||
|
if new_provider.get_name() == old_provider.get_name():
|
||||||
|
matching_new_provider = new_provider
|
||||||
|
break
|
||||||
|
|
||||||
|
if matching_new_provider:
|
||||||
|
# Transfer cumulative statistics
|
||||||
|
matching_new_provider.total_requests += old_provider.total_requests
|
||||||
|
matching_new_provider.successful_requests += old_provider.successful_requests
|
||||||
|
matching_new_provider.failed_requests += old_provider.failed_requests
|
||||||
|
matching_new_provider.total_relationships_found += old_provider.total_relationships_found
|
||||||
|
|
||||||
|
# Transfer cache statistics if available
|
||||||
|
if hasattr(old_provider, 'cache_hits'):
|
||||||
|
matching_new_provider.cache_hits += getattr(old_provider, 'cache_hits', 0)
|
||||||
|
matching_new_provider.cache_misses += getattr(old_provider, 'cache_misses', 0)
|
||||||
|
|
||||||
|
print(f"Consolidated {old_provider.get_name()} provider stats: {old_provider.total_requests} requests")
|
||||||
|
|
||||||
|
return new_scanner
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
print(f"Warning: Error during session consolidation: {e}")
|
||||||
|
return new_scanner
|
||||||
|
|
||||||
|
|
||||||
class SessionManager:
|
class SessionManager:
|
||||||
"""
|
"""
|
||||||
Manages multiple scanner instances for concurrent user sessions using Redis.
|
Manages single scanner session per user using Redis with user identification.
|
||||||
|
Enforces one active session per user for consistent state management.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, session_timeout_minutes: int = 0):
|
def __init__(self, session_timeout_minutes: int = 60):
|
||||||
"""
|
"""
|
||||||
Initialize session manager with a Redis backend.
|
Initialize session manager with Redis backend and user tracking.
|
||||||
"""
|
"""
|
||||||
if session_timeout_minutes is None:
|
|
||||||
session_timeout_minutes = config.session_timeout_minutes
|
|
||||||
|
|
||||||
self.redis_client = redis.StrictRedis(db=0, decode_responses=False)
|
self.redis_client = redis.StrictRedis(db=0, decode_responses=False)
|
||||||
self.session_timeout = session_timeout_minutes * 60 # Convert to seconds
|
self.session_timeout = session_timeout_minutes * 60 # Convert to seconds
|
||||||
self.lock = threading.Lock() # Lock for local operations, Redis handles atomic ops
|
self.lock = threading.Lock()
|
||||||
|
|
||||||
|
# User identification helper
|
||||||
|
self.user_identifier = UserIdentifier()
|
||||||
|
self.consolidator = SessionConsolidator()
|
||||||
|
|
||||||
# Start cleanup thread
|
# Start cleanup thread
|
||||||
self.cleanup_thread = threading.Thread(target=self._cleanup_loop, daemon=True)
|
self.cleanup_thread = threading.Thread(target=self._cleanup_loop, daemon=True)
|
||||||
self.cleanup_thread.start()
|
self.cleanup_thread.start()
|
||||||
|
|
||||||
print(f"SessionManager initialized with Redis backend and {session_timeout_minutes}min timeout")
|
print(f"SessionManager initialized with Redis backend, user tracking, and {session_timeout_minutes}min timeout")
|
||||||
|
|
||||||
def __getstate__(self):
|
def __getstate__(self):
|
||||||
"""Prepare SessionManager for pickling."""
|
"""Prepare SessionManager for pickling."""
|
||||||
state = self.__dict__.copy()
|
state = self.__dict__.copy()
|
||||||
# Exclude unpickleable attributes - Redis client and threading objects
|
# Exclude unpickleable attributes
|
||||||
unpicklable_attrs = ['lock', 'cleanup_thread', 'redis_client']
|
unpicklable_attrs = ['lock', 'cleanup_thread', 'redis_client']
|
||||||
for attr in unpicklable_attrs:
|
for attr in unpicklable_attrs:
|
||||||
if attr in state:
|
if attr in state:
|
||||||
@@ -57,67 +169,108 @@ class SessionManager:
|
|||||||
self.cleanup_thread.start()
|
self.cleanup_thread.start()
|
||||||
|
|
||||||
def _get_session_key(self, session_id: str) -> str:
|
def _get_session_key(self, session_id: str) -> str:
|
||||||
"""Generates the Redis key for a session."""
|
"""Generate Redis key for a session."""
|
||||||
return f"dnsrecon:session:{session_id}"
|
return f"dnsrecon:session:{session_id}"
|
||||||
|
|
||||||
|
def _get_user_session_key(self, user_fingerprint: str) -> str:
|
||||||
|
"""Generate Redis key for user -> session mapping."""
|
||||||
|
return f"dnsrecon:user:{user_fingerprint}"
|
||||||
|
|
||||||
def _get_stop_signal_key(self, session_id: str) -> str:
|
def _get_stop_signal_key(self, session_id: str) -> str:
|
||||||
"""Generates the Redis key for a session's stop signal."""
|
"""Generate Redis key for session stop signal."""
|
||||||
return f"dnsrecon:stop:{session_id}"
|
return f"dnsrecon:stop:{session_id}"
|
||||||
|
|
||||||
def create_session(self) -> str:
|
def create_or_replace_user_session(self, client_ip: str, user_agent: str) -> str:
|
||||||
"""
|
"""
|
||||||
Create a new user session and store it in Redis.
|
Create new session for user, replacing any existing session.
|
||||||
|
Consolidates data from previous session if it exists.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
client_ip: Client IP address
|
||||||
|
user_agent: User-Agent header
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
New session ID
|
||||||
"""
|
"""
|
||||||
session_id = str(uuid.uuid4())
|
user_fingerprint = self.user_identifier.generate_user_fingerprint(client_ip, user_agent)
|
||||||
print(f"=== CREATING SESSION {session_id} IN REDIS ===")
|
new_session_id = str(uuid.uuid4())
|
||||||
|
|
||||||
|
print(f"=== CREATING/REPLACING SESSION FOR USER {user_fingerprint} ===")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
# Check for existing user session
|
||||||
|
existing_session_id = self._get_user_current_session(user_fingerprint)
|
||||||
|
old_scanner = None
|
||||||
|
|
||||||
|
if existing_session_id:
|
||||||
|
print(f"Found existing session {existing_session_id} for user {user_fingerprint}")
|
||||||
|
# Get old scanner data for consolidation
|
||||||
|
old_scanner = self.get_session(existing_session_id)
|
||||||
|
# Terminate old session
|
||||||
|
self._terminate_session_internal(existing_session_id, cleanup_user_mapping=False)
|
||||||
|
print(f"Terminated old session {existing_session_id}")
|
||||||
|
|
||||||
|
# Create new session config and scanner
|
||||||
from core.session_config import create_session_config
|
from core.session_config import create_session_config
|
||||||
session_config = create_session_config()
|
session_config = create_session_config()
|
||||||
scanner_instance = Scanner(session_config=session_config)
|
new_scanner = Scanner(session_config=session_config)
|
||||||
|
|
||||||
# Set the session ID on the scanner for cross-process stop signal management
|
# Set session ID on scanner for cross-process operations
|
||||||
scanner_instance.session_id = session_id
|
new_scanner.session_id = new_session_id
|
||||||
|
|
||||||
|
# Consolidate data from old session if available
|
||||||
|
if old_scanner:
|
||||||
|
new_scanner = self.consolidator.consolidate_scanner_data(old_scanner, new_scanner)
|
||||||
|
print(f"Consolidated data from previous session")
|
||||||
|
|
||||||
|
# Create session data
|
||||||
session_data = {
|
session_data = {
|
||||||
'scanner': scanner_instance,
|
'scanner': new_scanner,
|
||||||
'config': session_config,
|
'config': session_config,
|
||||||
'created_at': time.time(),
|
'created_at': time.time(),
|
||||||
'last_activity': time.time(),
|
'last_activity': time.time(),
|
||||||
'status': 'active'
|
'status': 'active',
|
||||||
|
'user_fingerprint': user_fingerprint,
|
||||||
|
'client_ip': client_ip,
|
||||||
|
'user_agent': user_agent[:200] # Truncate for storage
|
||||||
}
|
}
|
||||||
|
|
||||||
# Serialize the entire session data dictionary using pickle
|
# Store session in Redis
|
||||||
|
session_key = self._get_session_key(new_session_id)
|
||||||
serialized_data = pickle.dumps(session_data)
|
serialized_data = pickle.dumps(session_data)
|
||||||
|
|
||||||
# Store in Redis
|
|
||||||
session_key = self._get_session_key(session_id)
|
|
||||||
self.redis_client.setex(session_key, self.session_timeout, serialized_data)
|
self.redis_client.setex(session_key, self.session_timeout, serialized_data)
|
||||||
|
|
||||||
# Initialize stop signal as False
|
# Update user -> session mapping
|
||||||
stop_key = self._get_stop_signal_key(session_id)
|
user_session_key = self._get_user_session_key(user_fingerprint)
|
||||||
|
self.redis_client.setex(user_session_key, self.session_timeout, new_session_id.encode('utf-8'))
|
||||||
|
|
||||||
|
# Initialize stop signal
|
||||||
|
stop_key = self._get_stop_signal_key(new_session_id)
|
||||||
self.redis_client.setex(stop_key, self.session_timeout, b'0')
|
self.redis_client.setex(stop_key, self.session_timeout, b'0')
|
||||||
|
|
||||||
print(f"Session {session_id} stored in Redis with stop signal initialized")
|
print(f"Created new session {new_session_id} for user {user_fingerprint}")
|
||||||
return session_id
|
return new_session_id
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f"ERROR: Failed to create session {session_id}: {e}")
|
print(f"ERROR: Failed to create session for user {user_fingerprint}: {e}")
|
||||||
raise
|
raise
|
||||||
|
|
||||||
|
def _get_user_current_session(self, user_fingerprint: str) -> Optional[str]:
|
||||||
|
"""Get current session ID for a user."""
|
||||||
|
try:
|
||||||
|
user_session_key = self._get_user_session_key(user_fingerprint)
|
||||||
|
session_id_bytes = self.redis_client.get(user_session_key)
|
||||||
|
if session_id_bytes:
|
||||||
|
return session_id_bytes.decode('utf-8')
|
||||||
|
return None
|
||||||
|
except Exception as e:
|
||||||
|
print(f"Error getting user session: {e}")
|
||||||
|
return None
|
||||||
|
|
||||||
def set_stop_signal(self, session_id: str) -> bool:
|
def set_stop_signal(self, session_id: str) -> bool:
|
||||||
"""
|
"""Set stop signal for session (cross-process safe)."""
|
||||||
Set the stop signal for a session (cross-process safe).
|
|
||||||
|
|
||||||
Args:
|
|
||||||
session_id: Session identifier
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
bool: True if signal was set successfully
|
|
||||||
"""
|
|
||||||
try:
|
try:
|
||||||
stop_key = self._get_stop_signal_key(session_id)
|
stop_key = self._get_stop_signal_key(session_id)
|
||||||
# Set stop signal to '1' with the same TTL as the session
|
|
||||||
self.redis_client.setex(stop_key, self.session_timeout, b'1')
|
self.redis_client.setex(stop_key, self.session_timeout, b'1')
|
||||||
print(f"Stop signal set for session {session_id}")
|
print(f"Stop signal set for session {session_id}")
|
||||||
return True
|
return True
|
||||||
@@ -126,15 +279,7 @@ class SessionManager:
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def is_stop_requested(self, session_id: str) -> bool:
|
def is_stop_requested(self, session_id: str) -> bool:
|
||||||
"""
|
"""Check if stop is requested for session (cross-process safe)."""
|
||||||
Check if stop is requested for a session (cross-process safe).
|
|
||||||
|
|
||||||
Args:
|
|
||||||
session_id: Session identifier
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
bool: True if stop is requested
|
|
||||||
"""
|
|
||||||
try:
|
try:
|
||||||
stop_key = self._get_stop_signal_key(session_id)
|
stop_key = self._get_stop_signal_key(session_id)
|
||||||
value = self.redis_client.get(stop_key)
|
value = self.redis_client.get(stop_key)
|
||||||
@@ -144,15 +289,7 @@ class SessionManager:
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def clear_stop_signal(self, session_id: str) -> bool:
|
def clear_stop_signal(self, session_id: str) -> bool:
|
||||||
"""
|
"""Clear stop signal for session."""
|
||||||
Clear the stop signal for a session.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
session_id: Session identifier
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
bool: True if signal was cleared successfully
|
|
||||||
"""
|
|
||||||
try:
|
try:
|
||||||
stop_key = self._get_stop_signal_key(session_id)
|
stop_key = self._get_stop_signal_key(session_id)
|
||||||
self.redis_client.setex(stop_key, self.session_timeout, b'0')
|
self.redis_client.setex(stop_key, self.session_timeout, b'0')
|
||||||
@@ -163,13 +300,13 @@ class SessionManager:
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def _get_session_data(self, session_id: str) -> Optional[Dict[str, Any]]:
|
def _get_session_data(self, session_id: str) -> Optional[Dict[str, Any]]:
|
||||||
"""Retrieves and deserializes session data from Redis."""
|
"""Retrieve and deserialize session data from Redis."""
|
||||||
try:
|
try:
|
||||||
session_key = self._get_session_key(session_id)
|
session_key = self._get_session_key(session_id)
|
||||||
serialized_data = self.redis_client.get(session_key)
|
serialized_data = self.redis_client.get(session_key)
|
||||||
if serialized_data:
|
if serialized_data:
|
||||||
session_data = pickle.loads(serialized_data)
|
session_data = pickle.loads(serialized_data)
|
||||||
# Ensure the scanner has the correct session ID for stop signal checking
|
# Ensure scanner has correct session ID
|
||||||
if 'scanner' in session_data and session_data['scanner']:
|
if 'scanner' in session_data and session_data['scanner']:
|
||||||
session_data['scanner'].session_id = session_id
|
session_data['scanner'].session_id = session_id
|
||||||
return session_data
|
return session_data
|
||||||
@@ -179,37 +316,32 @@ class SessionManager:
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
def _save_session_data(self, session_id: str, session_data: Dict[str, Any]) -> bool:
|
def _save_session_data(self, session_id: str, session_data: Dict[str, Any]) -> bool:
|
||||||
"""
|
"""Serialize and save session data to Redis with updated TTL."""
|
||||||
Serializes and saves session data back to Redis with updated TTL.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
bool: True if save was successful
|
|
||||||
"""
|
|
||||||
try:
|
try:
|
||||||
session_key = self._get_session_key(session_id)
|
session_key = self._get_session_key(session_id)
|
||||||
serialized_data = pickle.dumps(session_data)
|
serialized_data = pickle.dumps(session_data)
|
||||||
result = self.redis_client.setex(session_key, self.session_timeout, serialized_data)
|
result = self.redis_client.setex(session_key, self.session_timeout, serialized_data)
|
||||||
|
|
||||||
|
# Also refresh user mapping TTL if available
|
||||||
|
if 'user_fingerprint' in session_data:
|
||||||
|
user_session_key = self._get_user_session_key(session_data['user_fingerprint'])
|
||||||
|
self.redis_client.setex(user_session_key, self.session_timeout, session_id.encode('utf-8'))
|
||||||
|
|
||||||
return result
|
return result
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f"ERROR: Failed to save session data for {session_id}: {e}")
|
print(f"ERROR: Failed to save session data for {session_id}: {e}")
|
||||||
return False
|
return False
|
||||||
|
|
||||||
def update_session_scanner(self, session_id: str, scanner: 'Scanner') -> bool:
|
def update_session_scanner(self, session_id: str, scanner: 'Scanner') -> bool:
|
||||||
"""
|
"""Update scanner object in session with immediate persistence."""
|
||||||
Updates just the scanner object in a session with immediate persistence.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
bool: True if update was successful
|
|
||||||
"""
|
|
||||||
try:
|
try:
|
||||||
session_data = self._get_session_data(session_id)
|
session_data = self._get_session_data(session_id)
|
||||||
if session_data:
|
if session_data:
|
||||||
# Ensure scanner has the session ID
|
# Ensure scanner has session ID
|
||||||
scanner.session_id = session_id
|
scanner.session_id = session_id
|
||||||
session_data['scanner'] = scanner
|
session_data['scanner'] = scanner
|
||||||
session_data['last_activity'] = time.time()
|
session_data['last_activity'] = time.time()
|
||||||
|
|
||||||
# Immediately save to Redis for GUI updates
|
|
||||||
success = self._save_session_data(session_id, session_data)
|
success = self._save_session_data(session_id, session_data)
|
||||||
if success:
|
if success:
|
||||||
print(f"Scanner state updated for session {session_id} (status: {scanner.status})")
|
print(f"Scanner state updated for session {session_id} (status: {scanner.status})")
|
||||||
@@ -224,16 +356,7 @@ class SessionManager:
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def update_scanner_status(self, session_id: str, status: str) -> bool:
|
def update_scanner_status(self, session_id: str, status: str) -> bool:
|
||||||
"""
|
"""Quickly update scanner status for immediate GUI feedback."""
|
||||||
Quickly update just the scanner status for immediate GUI feedback.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
session_id: Session identifier
|
|
||||||
status: New scanner status
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
bool: True if update was successful
|
|
||||||
"""
|
|
||||||
try:
|
try:
|
||||||
session_data = self._get_session_data(session_id)
|
session_data = self._get_session_data(session_id)
|
||||||
if session_data and 'scanner' in session_data:
|
if session_data and 'scanner' in session_data:
|
||||||
@@ -252,9 +375,7 @@ class SessionManager:
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def get_session(self, session_id: str) -> Optional[Scanner]:
|
def get_session(self, session_id: str) -> Optional[Scanner]:
|
||||||
"""
|
"""Get scanner instance for session with session ID management."""
|
||||||
Get scanner instance for a session from Redis with session ID management.
|
|
||||||
"""
|
|
||||||
if not session_id:
|
if not session_id:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
@@ -269,21 +390,13 @@ class SessionManager:
|
|||||||
|
|
||||||
scanner = session_data.get('scanner')
|
scanner = session_data.get('scanner')
|
||||||
if scanner:
|
if scanner:
|
||||||
# Ensure the scanner can check the Redis-based stop signal
|
# Ensure scanner can check Redis-based stop signal
|
||||||
scanner.session_id = session_id
|
scanner.session_id = session_id
|
||||||
|
|
||||||
return scanner
|
return scanner
|
||||||
|
|
||||||
def get_session_status_only(self, session_id: str) -> Optional[str]:
|
def get_session_status_only(self, session_id: str) -> Optional[str]:
|
||||||
"""
|
"""Get scanner status without full session retrieval (for performance)."""
|
||||||
Get just the scanner status without full session retrieval (for performance).
|
|
||||||
|
|
||||||
Args:
|
|
||||||
session_id: Session identifier
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Scanner status string or None if not found
|
|
||||||
"""
|
|
||||||
try:
|
try:
|
||||||
session_data = self._get_session_data(session_id)
|
session_data = self._get_session_data(session_id)
|
||||||
if session_data and 'scanner' in session_data:
|
if session_data and 'scanner' in session_data:
|
||||||
@@ -294,16 +407,18 @@ class SessionManager:
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
def terminate_session(self, session_id: str) -> bool:
|
def terminate_session(self, session_id: str) -> bool:
|
||||||
"""
|
"""Terminate specific session with reliable stop signal and immediate status update."""
|
||||||
Terminate a specific session in Redis with reliable stop signal and immediate status update.
|
return self._terminate_session_internal(session_id, cleanup_user_mapping=True)
|
||||||
"""
|
|
||||||
|
def _terminate_session_internal(self, session_id: str, cleanup_user_mapping: bool = True) -> bool:
|
||||||
|
"""Internal session termination with configurable user mapping cleanup."""
|
||||||
print(f"=== TERMINATING SESSION {session_id} ===")
|
print(f"=== TERMINATING SESSION {session_id} ===")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# First, set the stop signal
|
# Set stop signal first
|
||||||
self.set_stop_signal(session_id)
|
self.set_stop_signal(session_id)
|
||||||
|
|
||||||
# Update scanner status to stopped immediately for GUI feedback
|
# Update scanner status immediately for GUI feedback
|
||||||
self.update_scanner_status(session_id, 'stopped')
|
self.update_scanner_status(session_id, 'stopped')
|
||||||
|
|
||||||
session_data = self._get_session_data(session_id)
|
session_data = self._get_session_data(session_id)
|
||||||
@@ -314,16 +429,19 @@ class SessionManager:
|
|||||||
scanner = session_data.get('scanner')
|
scanner = session_data.get('scanner')
|
||||||
if scanner and scanner.status == 'running':
|
if scanner and scanner.status == 'running':
|
||||||
print(f"Stopping scan for session: {session_id}")
|
print(f"Stopping scan for session: {session_id}")
|
||||||
# The scanner will check the Redis stop signal
|
|
||||||
scanner.stop_scan()
|
scanner.stop_scan()
|
||||||
|
|
||||||
# Update the scanner state immediately
|
|
||||||
self.update_session_scanner(session_id, scanner)
|
self.update_session_scanner(session_id, scanner)
|
||||||
|
|
||||||
# Wait a moment for graceful shutdown
|
# Wait for graceful shutdown
|
||||||
time.sleep(0.5)
|
time.sleep(0.5)
|
||||||
|
|
||||||
# Delete session data and stop signal from Redis
|
# Clean up user mapping if requested
|
||||||
|
if cleanup_user_mapping and 'user_fingerprint' in session_data:
|
||||||
|
user_session_key = self._get_user_session_key(session_data['user_fingerprint'])
|
||||||
|
self.redis_client.delete(user_session_key)
|
||||||
|
print(f"Cleaned up user mapping for {session_data['user_fingerprint']}")
|
||||||
|
|
||||||
|
# Delete session data and stop signal
|
||||||
session_key = self._get_session_key(session_id)
|
session_key = self._get_session_key(session_id)
|
||||||
stop_key = self._get_stop_signal_key(session_id)
|
stop_key = self._get_stop_signal_key(session_id)
|
||||||
self.redis_client.delete(session_key)
|
self.redis_client.delete(session_key)
|
||||||
@@ -337,35 +455,72 @@ class SessionManager:
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def _cleanup_loop(self) -> None:
|
def _cleanup_loop(self) -> None:
|
||||||
"""
|
"""Background thread to cleanup inactive sessions and orphaned signals."""
|
||||||
Background thread to cleanup inactive sessions and orphaned stop signals.
|
|
||||||
"""
|
|
||||||
while True:
|
while True:
|
||||||
try:
|
try:
|
||||||
# Clean up orphaned stop signals
|
# Clean up orphaned stop signals
|
||||||
stop_keys = self.redis_client.keys("dnsrecon:stop:*")
|
stop_keys = self.redis_client.keys("dnsrecon:stop:*")
|
||||||
for stop_key in stop_keys:
|
for stop_key in stop_keys:
|
||||||
# Extract session ID from stop key
|
|
||||||
session_id = stop_key.decode('utf-8').split(':')[-1]
|
session_id = stop_key.decode('utf-8').split(':')[-1]
|
||||||
session_key = self._get_session_key(session_id)
|
session_key = self._get_session_key(session_id)
|
||||||
|
|
||||||
# If session doesn't exist but stop signal does, clean it up
|
|
||||||
if not self.redis_client.exists(session_key):
|
if not self.redis_client.exists(session_key):
|
||||||
self.redis_client.delete(stop_key)
|
self.redis_client.delete(stop_key)
|
||||||
print(f"Cleaned up orphaned stop signal for session {session_id}")
|
print(f"Cleaned up orphaned stop signal for session {session_id}")
|
||||||
|
|
||||||
|
# Clean up orphaned user mappings
|
||||||
|
user_keys = self.redis_client.keys("dnsrecon:user:*")
|
||||||
|
for user_key in user_keys:
|
||||||
|
session_id_bytes = self.redis_client.get(user_key)
|
||||||
|
if session_id_bytes:
|
||||||
|
session_id = session_id_bytes.decode('utf-8')
|
||||||
|
session_key = self._get_session_key(session_id)
|
||||||
|
|
||||||
|
if not self.redis_client.exists(session_key):
|
||||||
|
self.redis_client.delete(user_key)
|
||||||
|
print(f"Cleaned up orphaned user mapping for session {session_id}")
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f"Error in cleanup loop: {e}")
|
print(f"Error in cleanup loop: {e}")
|
||||||
|
|
||||||
time.sleep(300) # Sleep for 5 minutes
|
time.sleep(300) # Sleep for 5 minutes
|
||||||
|
|
||||||
|
def list_active_sessions(self) -> List[Dict[str, Any]]:
|
||||||
|
"""List all active sessions for admin purposes."""
|
||||||
|
try:
|
||||||
|
session_keys = self.redis_client.keys("dnsrecon:session:*")
|
||||||
|
sessions = []
|
||||||
|
|
||||||
|
for session_key in session_keys:
|
||||||
|
session_id = session_key.decode('utf-8').split(':')[-1]
|
||||||
|
session_data = self._get_session_data(session_id)
|
||||||
|
|
||||||
|
if session_data:
|
||||||
|
scanner = session_data.get('scanner')
|
||||||
|
sessions.append({
|
||||||
|
'session_id': session_id,
|
||||||
|
'user_fingerprint': session_data.get('user_fingerprint', 'unknown'),
|
||||||
|
'client_ip': session_data.get('client_ip', 'unknown'),
|
||||||
|
'created_at': session_data.get('created_at'),
|
||||||
|
'last_activity': session_data.get('last_activity'),
|
||||||
|
'scanner_status': scanner.status if scanner else 'unknown',
|
||||||
|
'current_target': scanner.current_target if scanner else None
|
||||||
|
})
|
||||||
|
|
||||||
|
return sessions
|
||||||
|
except Exception as e:
|
||||||
|
print(f"ERROR: Failed to list active sessions: {e}")
|
||||||
|
return []
|
||||||
|
|
||||||
def get_statistics(self) -> Dict[str, Any]:
|
def get_statistics(self) -> Dict[str, Any]:
|
||||||
"""Get session manager statistics."""
|
"""Get session manager statistics."""
|
||||||
try:
|
try:
|
||||||
session_keys = self.redis_client.keys("dnsrecon:session:*")
|
session_keys = self.redis_client.keys("dnsrecon:session:*")
|
||||||
|
user_keys = self.redis_client.keys("dnsrecon:user:*")
|
||||||
stop_keys = self.redis_client.keys("dnsrecon:stop:*")
|
stop_keys = self.redis_client.keys("dnsrecon:stop:*")
|
||||||
|
|
||||||
active_sessions = len(session_keys)
|
active_sessions = len(session_keys)
|
||||||
|
unique_users = len(user_keys)
|
||||||
running_scans = 0
|
running_scans = 0
|
||||||
|
|
||||||
for session_key in session_keys:
|
for session_key in session_keys:
|
||||||
@@ -376,16 +531,46 @@ class SessionManager:
|
|||||||
|
|
||||||
return {
|
return {
|
||||||
'total_active_sessions': active_sessions,
|
'total_active_sessions': active_sessions,
|
||||||
|
'unique_users': unique_users,
|
||||||
'running_scans': running_scans,
|
'running_scans': running_scans,
|
||||||
'total_stop_signals': len(stop_keys)
|
'total_stop_signals': len(stop_keys),
|
||||||
|
'average_sessions_per_user': round(active_sessions / unique_users, 2) if unique_users > 0 else 0
|
||||||
}
|
}
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f"ERROR: Failed to get statistics: {e}")
|
print(f"ERROR: Failed to get statistics: {e}")
|
||||||
return {
|
return {
|
||||||
'total_active_sessions': 0,
|
'total_active_sessions': 0,
|
||||||
|
'unique_users': 0,
|
||||||
'running_scans': 0,
|
'running_scans': 0,
|
||||||
'total_stop_signals': 0
|
'total_stop_signals': 0,
|
||||||
|
'average_sessions_per_user': 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
def get_session_info(self, session_id: str) -> Dict[str, Any]:
|
||||||
|
"""Get detailed information about a specific session."""
|
||||||
|
try:
|
||||||
|
session_data = self._get_session_data(session_id)
|
||||||
|
if not session_data:
|
||||||
|
return {'error': 'Session not found'}
|
||||||
|
|
||||||
|
scanner = session_data.get('scanner')
|
||||||
|
|
||||||
|
return {
|
||||||
|
'session_id': session_id,
|
||||||
|
'user_fingerprint': session_data.get('user_fingerprint', 'unknown'),
|
||||||
|
'client_ip': session_data.get('client_ip', 'unknown'),
|
||||||
|
'user_agent': session_data.get('user_agent', 'unknown'),
|
||||||
|
'created_at': session_data.get('created_at'),
|
||||||
|
'last_activity': session_data.get('last_activity'),
|
||||||
|
'status': session_data.get('status'),
|
||||||
|
'scanner_status': scanner.status if scanner else 'unknown',
|
||||||
|
'current_target': scanner.current_target if scanner else None,
|
||||||
|
'session_age_minutes': round((time.time() - session_data.get('created_at', time.time())) / 60, 1)
|
||||||
|
}
|
||||||
|
except Exception as e:
|
||||||
|
print(f"ERROR: Failed to get session info for {session_id}: {e}")
|
||||||
|
return {'error': f'Failed to get session info: {str(e)}'}
|
||||||
|
|
||||||
|
|
||||||
# Global session manager instance
|
# Global session manager instance
|
||||||
session_manager = SessionManager(session_timeout_minutes=60)
|
session_manager = SessionManager(session_timeout_minutes=60)
|
||||||
564
core/task_manager.py
Normal file
564
core/task_manager.py
Normal file
@@ -0,0 +1,564 @@
|
|||||||
|
# dnsrecon/core/task_manager.py
|
||||||
|
|
||||||
|
import threading
|
||||||
|
import time
|
||||||
|
import uuid
|
||||||
|
from enum import Enum
|
||||||
|
from dataclasses import dataclass, field
|
||||||
|
from typing import Dict, List, Optional, Any, Set
|
||||||
|
from datetime import datetime, timezone, timedelta
|
||||||
|
from collections import deque
|
||||||
|
|
||||||
|
from utils.helpers import _is_valid_ip, _is_valid_domain
|
||||||
|
|
||||||
|
|
||||||
|
class TaskStatus(Enum):
|
||||||
|
"""Enumeration of task execution statuses."""
|
||||||
|
PENDING = "pending"
|
||||||
|
RUNNING = "running"
|
||||||
|
SUCCEEDED = "succeeded"
|
||||||
|
FAILED_RETRYING = "failed_retrying"
|
||||||
|
FAILED_PERMANENT = "failed_permanent"
|
||||||
|
CANCELLED = "cancelled"
|
||||||
|
|
||||||
|
|
||||||
|
class TaskType(Enum):
|
||||||
|
"""Enumeration of task types for provider queries."""
|
||||||
|
DOMAIN_QUERY = "domain_query"
|
||||||
|
IP_QUERY = "ip_query"
|
||||||
|
GRAPH_UPDATE = "graph_update"
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass
|
||||||
|
class TaskResult:
|
||||||
|
"""Result of a task execution."""
|
||||||
|
success: bool
|
||||||
|
data: Optional[Any] = None
|
||||||
|
error: Optional[str] = None
|
||||||
|
metadata: Dict[str, Any] = field(default_factory=dict)
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass
|
||||||
|
class ReconTask:
|
||||||
|
"""Represents a single reconnaissance task with retry logic."""
|
||||||
|
task_id: str
|
||||||
|
task_type: TaskType
|
||||||
|
target: str
|
||||||
|
provider_name: str
|
||||||
|
depth: int
|
||||||
|
status: TaskStatus = TaskStatus.PENDING
|
||||||
|
created_at: datetime = field(default_factory=lambda: datetime.now(timezone.utc))
|
||||||
|
|
||||||
|
# Retry configuration
|
||||||
|
max_retries: int = 3
|
||||||
|
current_retry: int = 0
|
||||||
|
base_backoff_seconds: float = 1.0
|
||||||
|
max_backoff_seconds: float = 60.0
|
||||||
|
|
||||||
|
# Execution tracking
|
||||||
|
last_attempt_at: Optional[datetime] = None
|
||||||
|
next_retry_at: Optional[datetime] = None
|
||||||
|
execution_history: List[Dict[str, Any]] = field(default_factory=list)
|
||||||
|
|
||||||
|
# Results
|
||||||
|
result: Optional[TaskResult] = None
|
||||||
|
|
||||||
|
def __post_init__(self):
|
||||||
|
"""Initialize additional fields after creation."""
|
||||||
|
if not self.task_id:
|
||||||
|
self.task_id = str(uuid.uuid4())[:8]
|
||||||
|
|
||||||
|
def calculate_next_retry_time(self) -> datetime:
|
||||||
|
"""Calculate next retry time with exponential backoff and jitter."""
|
||||||
|
if self.current_retry >= self.max_retries:
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Exponential backoff with jitter
|
||||||
|
backoff_time = min(
|
||||||
|
self.max_backoff_seconds,
|
||||||
|
self.base_backoff_seconds * (2 ** self.current_retry)
|
||||||
|
)
|
||||||
|
|
||||||
|
# Add jitter (±25%)
|
||||||
|
jitter = backoff_time * 0.25 * (0.5 - hash(self.task_id) % 1000 / 1000.0)
|
||||||
|
final_backoff = max(self.base_backoff_seconds, backoff_time + jitter)
|
||||||
|
|
||||||
|
return datetime.now(timezone.utc) + timedelta(seconds=final_backoff)
|
||||||
|
|
||||||
|
def should_retry(self) -> bool:
|
||||||
|
"""Determine if task should be retried based on status and retry count."""
|
||||||
|
if self.status != TaskStatus.FAILED_RETRYING:
|
||||||
|
return False
|
||||||
|
if self.current_retry >= self.max_retries:
|
||||||
|
return False
|
||||||
|
if self.next_retry_at and datetime.now(timezone.utc) < self.next_retry_at:
|
||||||
|
return False
|
||||||
|
return True
|
||||||
|
|
||||||
|
def mark_failed(self, error: str, metadata: Dict[str, Any] = None):
|
||||||
|
"""Mark task as failed and prepare for retry or permanent failure."""
|
||||||
|
self.current_retry += 1
|
||||||
|
self.last_attempt_at = datetime.now(timezone.utc)
|
||||||
|
|
||||||
|
# Record execution history
|
||||||
|
execution_record = {
|
||||||
|
'attempt': self.current_retry,
|
||||||
|
'timestamp': self.last_attempt_at.isoformat(),
|
||||||
|
'error': error,
|
||||||
|
'metadata': metadata or {}
|
||||||
|
}
|
||||||
|
self.execution_history.append(execution_record)
|
||||||
|
|
||||||
|
if self.current_retry >= self.max_retries:
|
||||||
|
self.status = TaskStatus.FAILED_PERMANENT
|
||||||
|
self.result = TaskResult(success=False, error=f"Permanent failure after {self.max_retries} attempts: {error}")
|
||||||
|
else:
|
||||||
|
self.status = TaskStatus.FAILED_RETRYING
|
||||||
|
self.next_retry_at = self.calculate_next_retry_time()
|
||||||
|
|
||||||
|
def mark_succeeded(self, data: Any = None, metadata: Dict[str, Any] = None):
|
||||||
|
"""Mark task as successfully completed."""
|
||||||
|
self.status = TaskStatus.SUCCEEDED
|
||||||
|
self.last_attempt_at = datetime.now(timezone.utc)
|
||||||
|
self.result = TaskResult(success=True, data=data, metadata=metadata or {})
|
||||||
|
|
||||||
|
# Record successful execution
|
||||||
|
execution_record = {
|
||||||
|
'attempt': self.current_retry + 1,
|
||||||
|
'timestamp': self.last_attempt_at.isoformat(),
|
||||||
|
'success': True,
|
||||||
|
'metadata': metadata or {}
|
||||||
|
}
|
||||||
|
self.execution_history.append(execution_record)
|
||||||
|
|
||||||
|
def get_summary(self) -> Dict[str, Any]:
|
||||||
|
"""Get task summary for progress reporting."""
|
||||||
|
return {
|
||||||
|
'task_id': self.task_id,
|
||||||
|
'task_type': self.task_type.value,
|
||||||
|
'target': self.target,
|
||||||
|
'provider': self.provider_name,
|
||||||
|
'status': self.status.value,
|
||||||
|
'current_retry': self.current_retry,
|
||||||
|
'max_retries': self.max_retries,
|
||||||
|
'created_at': self.created_at.isoformat(),
|
||||||
|
'last_attempt_at': self.last_attempt_at.isoformat() if self.last_attempt_at else None,
|
||||||
|
'next_retry_at': self.next_retry_at.isoformat() if self.next_retry_at else None,
|
||||||
|
'total_attempts': len(self.execution_history),
|
||||||
|
'has_result': self.result is not None
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class TaskQueue:
|
||||||
|
"""Thread-safe task queue with retry logic and priority handling."""
|
||||||
|
|
||||||
|
def __init__(self, max_concurrent_tasks: int = 5):
|
||||||
|
"""Initialize task queue."""
|
||||||
|
self.max_concurrent_tasks = max_concurrent_tasks
|
||||||
|
self.tasks: Dict[str, ReconTask] = {}
|
||||||
|
self.pending_queue = deque()
|
||||||
|
self.retry_queue = deque()
|
||||||
|
self.running_tasks: Set[str] = set()
|
||||||
|
|
||||||
|
self._lock = threading.Lock()
|
||||||
|
self._stop_event = threading.Event()
|
||||||
|
|
||||||
|
def __getstate__(self):
|
||||||
|
"""Prepare TaskQueue for pickling by excluding unpicklable objects."""
|
||||||
|
state = self.__dict__.copy()
|
||||||
|
# Exclude the unpickleable '_lock' and '_stop_event' attributes
|
||||||
|
if '_lock' in state:
|
||||||
|
del state['_lock']
|
||||||
|
if '_stop_event' in state:
|
||||||
|
del state['_stop_event']
|
||||||
|
return state
|
||||||
|
|
||||||
|
def __setstate__(self, state):
|
||||||
|
"""Restore TaskQueue after unpickling by reconstructing threading objects."""
|
||||||
|
self.__dict__.update(state)
|
||||||
|
# Re-initialize the '_lock' and '_stop_event' attributes
|
||||||
|
self._lock = threading.Lock()
|
||||||
|
self._stop_event = threading.Event()
|
||||||
|
|
||||||
|
def add_task(self, task: ReconTask) -> str:
|
||||||
|
"""Add task to queue."""
|
||||||
|
with self._lock:
|
||||||
|
self.tasks[task.task_id] = task
|
||||||
|
self.pending_queue.append(task.task_id)
|
||||||
|
print(f"Added task {task.task_id}: {task.provider_name} query for {task.target}")
|
||||||
|
return task.task_id
|
||||||
|
|
||||||
|
def get_next_ready_task(self) -> Optional[ReconTask]:
|
||||||
|
"""Get next task ready for execution."""
|
||||||
|
with self._lock:
|
||||||
|
# Check if we have room for more concurrent tasks
|
||||||
|
if len(self.running_tasks) >= self.max_concurrent_tasks:
|
||||||
|
return None
|
||||||
|
|
||||||
|
# First priority: retry queue (tasks ready for retry)
|
||||||
|
while self.retry_queue:
|
||||||
|
task_id = self.retry_queue.popleft()
|
||||||
|
if task_id in self.tasks:
|
||||||
|
task = self.tasks[task_id]
|
||||||
|
if task.should_retry():
|
||||||
|
task.status = TaskStatus.RUNNING
|
||||||
|
self.running_tasks.add(task_id)
|
||||||
|
print(f"Retrying task {task_id} (attempt {task.current_retry + 1})")
|
||||||
|
return task
|
||||||
|
|
||||||
|
# Second priority: pending queue (new tasks)
|
||||||
|
while self.pending_queue:
|
||||||
|
task_id = self.pending_queue.popleft()
|
||||||
|
if task_id in self.tasks:
|
||||||
|
task = self.tasks[task_id]
|
||||||
|
if task.status == TaskStatus.PENDING:
|
||||||
|
task.status = TaskStatus.RUNNING
|
||||||
|
self.running_tasks.add(task_id)
|
||||||
|
print(f"Starting task {task_id}")
|
||||||
|
return task
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
def complete_task(self, task_id: str, success: bool, data: Any = None,
|
||||||
|
error: str = None, metadata: Dict[str, Any] = None):
|
||||||
|
"""Mark task as completed (success or failure)."""
|
||||||
|
with self._lock:
|
||||||
|
if task_id not in self.tasks:
|
||||||
|
return
|
||||||
|
|
||||||
|
task = self.tasks[task_id]
|
||||||
|
self.running_tasks.discard(task_id)
|
||||||
|
|
||||||
|
if success:
|
||||||
|
task.mark_succeeded(data=data, metadata=metadata)
|
||||||
|
print(f"Task {task_id} succeeded")
|
||||||
|
else:
|
||||||
|
task.mark_failed(error or "Unknown error", metadata=metadata)
|
||||||
|
if task.status == TaskStatus.FAILED_RETRYING:
|
||||||
|
self.retry_queue.append(task_id)
|
||||||
|
print(f"Task {task_id} failed, scheduled for retry at {task.next_retry_at}")
|
||||||
|
else:
|
||||||
|
print(f"Task {task_id} permanently failed after {task.current_retry} attempts")
|
||||||
|
|
||||||
|
def cancel_all_tasks(self):
|
||||||
|
"""Cancel all pending and running tasks."""
|
||||||
|
with self._lock:
|
||||||
|
self._stop_event.set()
|
||||||
|
for task in self.tasks.values():
|
||||||
|
if task.status in [TaskStatus.PENDING, TaskStatus.RUNNING, TaskStatus.FAILED_RETRYING]:
|
||||||
|
task.status = TaskStatus.CANCELLED
|
||||||
|
self.pending_queue.clear()
|
||||||
|
self.retry_queue.clear()
|
||||||
|
self.running_tasks.clear()
|
||||||
|
print("All tasks cancelled")
|
||||||
|
|
||||||
|
def is_complete(self) -> bool:
|
||||||
|
"""Check if all tasks are complete (succeeded, permanently failed, or cancelled)."""
|
||||||
|
with self._lock:
|
||||||
|
for task in self.tasks.values():
|
||||||
|
if task.status in [TaskStatus.PENDING, TaskStatus.RUNNING, TaskStatus.FAILED_RETRYING]:
|
||||||
|
return False
|
||||||
|
return True
|
||||||
|
|
||||||
|
def get_statistics(self) -> Dict[str, Any]:
|
||||||
|
"""Get queue statistics."""
|
||||||
|
with self._lock:
|
||||||
|
stats = {
|
||||||
|
'total_tasks': len(self.tasks),
|
||||||
|
'pending': len(self.pending_queue),
|
||||||
|
'running': len(self.running_tasks),
|
||||||
|
'retry_queue': len(self.retry_queue),
|
||||||
|
'succeeded': 0,
|
||||||
|
'failed_permanent': 0,
|
||||||
|
'cancelled': 0,
|
||||||
|
'failed_retrying': 0
|
||||||
|
}
|
||||||
|
|
||||||
|
for task in self.tasks.values():
|
||||||
|
if task.status == TaskStatus.SUCCEEDED:
|
||||||
|
stats['succeeded'] += 1
|
||||||
|
elif task.status == TaskStatus.FAILED_PERMANENT:
|
||||||
|
stats['failed_permanent'] += 1
|
||||||
|
elif task.status == TaskStatus.CANCELLED:
|
||||||
|
stats['cancelled'] += 1
|
||||||
|
elif task.status == TaskStatus.FAILED_RETRYING:
|
||||||
|
stats['failed_retrying'] += 1
|
||||||
|
|
||||||
|
stats['completion_rate'] = (stats['succeeded'] / stats['total_tasks'] * 100) if stats['total_tasks'] > 0 else 0
|
||||||
|
stats['is_complete'] = self.is_complete()
|
||||||
|
|
||||||
|
return stats
|
||||||
|
|
||||||
|
def get_task_summaries(self) -> List[Dict[str, Any]]:
|
||||||
|
"""Get summaries of all tasks for detailed progress reporting."""
|
||||||
|
with self._lock:
|
||||||
|
return [task.get_summary() for task in self.tasks.values()]
|
||||||
|
|
||||||
|
def get_failed_tasks(self) -> List[ReconTask]:
|
||||||
|
"""Get all permanently failed tasks for analysis."""
|
||||||
|
with self._lock:
|
||||||
|
return [task for task in self.tasks.values() if task.status == TaskStatus.FAILED_PERMANENT]
|
||||||
|
|
||||||
|
|
||||||
|
class TaskExecutor:
|
||||||
|
"""Executes reconnaissance tasks using providers."""
|
||||||
|
|
||||||
|
def __init__(self, providers: List, graph_manager, logger):
|
||||||
|
"""Initialize task executor."""
|
||||||
|
self.providers = {provider.get_name(): provider for provider in providers}
|
||||||
|
self.graph = graph_manager
|
||||||
|
self.logger = logger
|
||||||
|
|
||||||
|
def execute_task(self, task: ReconTask) -> TaskResult:
|
||||||
|
"""
|
||||||
|
Execute a single reconnaissance task.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
task: Task to execute
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
TaskResult with success/failure information
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
print(f"Executing task {task.task_id}: {task.provider_name} query for {task.target}")
|
||||||
|
|
||||||
|
provider = self.providers.get(task.provider_name)
|
||||||
|
if not provider:
|
||||||
|
return TaskResult(
|
||||||
|
success=False,
|
||||||
|
error=f"Provider {task.provider_name} not available"
|
||||||
|
)
|
||||||
|
|
||||||
|
if not provider.is_available():
|
||||||
|
return TaskResult(
|
||||||
|
success=False,
|
||||||
|
error=f"Provider {task.provider_name} is not available (missing API key or configuration)"
|
||||||
|
)
|
||||||
|
|
||||||
|
# Execute provider query based on task type
|
||||||
|
if task.task_type == TaskType.DOMAIN_QUERY:
|
||||||
|
if not _is_valid_domain(task.target):
|
||||||
|
return TaskResult(success=False, error=f"Invalid domain: {task.target}")
|
||||||
|
|
||||||
|
relationships = provider.query_domain(task.target)
|
||||||
|
|
||||||
|
elif task.task_type == TaskType.IP_QUERY:
|
||||||
|
if not _is_valid_ip(task.target):
|
||||||
|
return TaskResult(success=False, error=f"Invalid IP: {task.target}")
|
||||||
|
|
||||||
|
relationships = provider.query_ip(task.target)
|
||||||
|
|
||||||
|
else:
|
||||||
|
return TaskResult(success=False, error=f"Unsupported task type: {task.task_type}")
|
||||||
|
|
||||||
|
# Process results and update graph
|
||||||
|
new_targets = set()
|
||||||
|
relationships_added = 0
|
||||||
|
|
||||||
|
for source, target, rel_type, confidence, raw_data in relationships:
|
||||||
|
# Add nodes to graph
|
||||||
|
from core.graph_manager import NodeType
|
||||||
|
|
||||||
|
if _is_valid_ip(target):
|
||||||
|
self.graph.add_node(target, NodeType.IP)
|
||||||
|
new_targets.add(target)
|
||||||
|
elif target.startswith('AS') and target[2:].isdigit():
|
||||||
|
self.graph.add_node(target, NodeType.ASN)
|
||||||
|
elif _is_valid_domain(target):
|
||||||
|
self.graph.add_node(target, NodeType.DOMAIN)
|
||||||
|
new_targets.add(target)
|
||||||
|
|
||||||
|
# Add edge to graph
|
||||||
|
if self.graph.add_edge(source, target, rel_type, confidence, task.provider_name, raw_data):
|
||||||
|
relationships_added += 1
|
||||||
|
|
||||||
|
# Log forensic information
|
||||||
|
self.logger.logger.info(
|
||||||
|
f"Task {task.task_id} completed: {len(relationships)} relationships found, "
|
||||||
|
f"{relationships_added} added to graph, {len(new_targets)} new targets"
|
||||||
|
)
|
||||||
|
|
||||||
|
return TaskResult(
|
||||||
|
success=True,
|
||||||
|
data={
|
||||||
|
'relationships': relationships,
|
||||||
|
'new_targets': list(new_targets),
|
||||||
|
'relationships_added': relationships_added
|
||||||
|
},
|
||||||
|
metadata={
|
||||||
|
'provider': task.provider_name,
|
||||||
|
'target': task.target,
|
||||||
|
'depth': task.depth,
|
||||||
|
'execution_time': datetime.now(timezone.utc).isoformat()
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
error_msg = f"Task execution failed: {str(e)}"
|
||||||
|
print(f"ERROR: {error_msg} for task {task.task_id}")
|
||||||
|
self.logger.logger.error(error_msg)
|
||||||
|
|
||||||
|
return TaskResult(
|
||||||
|
success=False,
|
||||||
|
error=error_msg,
|
||||||
|
metadata={
|
||||||
|
'provider': task.provider_name,
|
||||||
|
'target': task.target,
|
||||||
|
'exception_type': type(e).__name__
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class TaskManager:
|
||||||
|
"""High-level task management for reconnaissance scans."""
|
||||||
|
|
||||||
|
def __init__(self, providers: List, graph_manager, logger, max_concurrent_tasks: int = 5):
|
||||||
|
"""Initialize task manager."""
|
||||||
|
self.task_queue = TaskQueue(max_concurrent_tasks)
|
||||||
|
self.task_executor = TaskExecutor(providers, graph_manager, logger)
|
||||||
|
self.logger = logger
|
||||||
|
|
||||||
|
# Execution control
|
||||||
|
self._stop_event = threading.Event()
|
||||||
|
self._execution_threads: List[threading.Thread] = []
|
||||||
|
self._is_running = False
|
||||||
|
|
||||||
|
def create_provider_tasks(self, target: str, depth: int, providers: List) -> List[str]:
|
||||||
|
"""
|
||||||
|
Create tasks for querying all eligible providers for a target.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
target: Domain or IP to query
|
||||||
|
depth: Current recursion depth
|
||||||
|
providers: List of available providers
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
List of created task IDs
|
||||||
|
"""
|
||||||
|
task_ids = []
|
||||||
|
is_ip = _is_valid_ip(target)
|
||||||
|
target_key = 'ips' if is_ip else 'domains'
|
||||||
|
task_type = TaskType.IP_QUERY if is_ip else TaskType.DOMAIN_QUERY
|
||||||
|
|
||||||
|
for provider in providers:
|
||||||
|
if provider.get_eligibility().get(target_key) and provider.is_available():
|
||||||
|
task = ReconTask(
|
||||||
|
task_id=str(uuid.uuid4())[:8],
|
||||||
|
task_type=task_type,
|
||||||
|
target=target,
|
||||||
|
provider_name=provider.get_name(),
|
||||||
|
depth=depth,
|
||||||
|
max_retries=3 # Configure retries per task type/provider
|
||||||
|
)
|
||||||
|
|
||||||
|
task_id = self.task_queue.add_task(task)
|
||||||
|
task_ids.append(task_id)
|
||||||
|
|
||||||
|
return task_ids
|
||||||
|
|
||||||
|
def start_execution(self, max_workers: int = 3):
|
||||||
|
"""Start task execution with specified number of worker threads."""
|
||||||
|
if self._is_running:
|
||||||
|
print("Task execution already running")
|
||||||
|
return
|
||||||
|
|
||||||
|
self._is_running = True
|
||||||
|
self._stop_event.clear()
|
||||||
|
|
||||||
|
print(f"Starting task execution with {max_workers} workers")
|
||||||
|
|
||||||
|
for i in range(max_workers):
|
||||||
|
worker_thread = threading.Thread(
|
||||||
|
target=self._worker_loop,
|
||||||
|
name=f"TaskWorker-{i+1}",
|
||||||
|
daemon=True
|
||||||
|
)
|
||||||
|
worker_thread.start()
|
||||||
|
self._execution_threads.append(worker_thread)
|
||||||
|
|
||||||
|
def stop_execution(self):
|
||||||
|
"""Stop task execution and cancel all tasks."""
|
||||||
|
print("Stopping task execution")
|
||||||
|
self._stop_event.set()
|
||||||
|
self.task_queue.cancel_all_tasks()
|
||||||
|
self._is_running = False
|
||||||
|
|
||||||
|
# Wait for worker threads to finish
|
||||||
|
for thread in self._execution_threads:
|
||||||
|
thread.join(timeout=5.0)
|
||||||
|
|
||||||
|
self._execution_threads.clear()
|
||||||
|
print("Task execution stopped")
|
||||||
|
|
||||||
|
def _worker_loop(self):
|
||||||
|
"""Worker thread loop for executing tasks."""
|
||||||
|
thread_name = threading.current_thread().name
|
||||||
|
print(f"{thread_name} started")
|
||||||
|
|
||||||
|
while not self._stop_event.is_set():
|
||||||
|
try:
|
||||||
|
# Get next task to execute
|
||||||
|
task = self.task_queue.get_next_ready_task()
|
||||||
|
|
||||||
|
if task is None:
|
||||||
|
# No tasks ready, check if we should exit
|
||||||
|
if self.task_queue.is_complete() or self._stop_event.is_set():
|
||||||
|
break
|
||||||
|
time.sleep(0.1) # Brief sleep before checking again
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Execute the task
|
||||||
|
result = self.task_executor.execute_task(task)
|
||||||
|
|
||||||
|
# Complete the task in queue
|
||||||
|
self.task_queue.complete_task(
|
||||||
|
task.task_id,
|
||||||
|
success=result.success,
|
||||||
|
data=result.data,
|
||||||
|
error=result.error,
|
||||||
|
metadata=result.metadata
|
||||||
|
)
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
print(f"ERROR: Worker {thread_name} encountered error: {e}")
|
||||||
|
# Continue running even if individual task fails
|
||||||
|
continue
|
||||||
|
|
||||||
|
print(f"{thread_name} finished")
|
||||||
|
|
||||||
|
def wait_for_completion(self, timeout_seconds: int = 300) -> bool:
|
||||||
|
"""
|
||||||
|
Wait for all tasks to complete.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
timeout_seconds: Maximum time to wait
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
True if all tasks completed, False if timeout
|
||||||
|
"""
|
||||||
|
start_time = time.time()
|
||||||
|
|
||||||
|
while time.time() - start_time < timeout_seconds:
|
||||||
|
if self.task_queue.is_complete():
|
||||||
|
return True
|
||||||
|
|
||||||
|
if self._stop_event.is_set():
|
||||||
|
return False
|
||||||
|
|
||||||
|
time.sleep(1.0) # Check every second
|
||||||
|
|
||||||
|
print(f"Timeout waiting for task completion after {timeout_seconds} seconds")
|
||||||
|
return False
|
||||||
|
|
||||||
|
def get_progress_report(self) -> Dict[str, Any]:
|
||||||
|
"""Get detailed progress report for UI updates."""
|
||||||
|
stats = self.task_queue.get_statistics()
|
||||||
|
failed_tasks = self.task_queue.get_failed_tasks()
|
||||||
|
|
||||||
|
return {
|
||||||
|
'statistics': stats,
|
||||||
|
'failed_tasks': [task.get_summary() for task in failed_tasks],
|
||||||
|
'is_running': self._is_running,
|
||||||
|
'worker_count': len(self._execution_threads),
|
||||||
|
'detailed_tasks': self.task_queue.get_task_summaries() if stats['total_tasks'] < 50 else [] # Limit detail for performance
|
||||||
|
}
|
||||||
@@ -3,14 +3,18 @@
|
|||||||
import time
|
import time
|
||||||
import requests
|
import requests
|
||||||
import threading
|
import threading
|
||||||
|
import os
|
||||||
|
import json
|
||||||
|
import hashlib
|
||||||
from abc import ABC, abstractmethod
|
from abc import ABC, abstractmethod
|
||||||
from typing import List, Dict, Any, Optional, Tuple
|
from typing import List, Dict, Any, Optional, Tuple
|
||||||
|
from datetime import datetime, timezone
|
||||||
|
|
||||||
from core.logger import get_forensic_logger
|
from core.logger import get_forensic_logger
|
||||||
|
|
||||||
|
|
||||||
class RateLimiter:
|
class RateLimiter:
|
||||||
"""Simple rate limiter for API calls."""
|
"""Thread-safe rate limiter for API calls."""
|
||||||
|
|
||||||
def __init__(self, requests_per_minute: int):
|
def __init__(self, requests_per_minute: int):
|
||||||
"""
|
"""
|
||||||
@@ -22,36 +26,152 @@ class RateLimiter:
|
|||||||
self.requests_per_minute = requests_per_minute
|
self.requests_per_minute = requests_per_minute
|
||||||
self.min_interval = 60.0 / requests_per_minute
|
self.min_interval = 60.0 / requests_per_minute
|
||||||
self.last_request_time = 0
|
self.last_request_time = 0
|
||||||
|
self._lock = threading.Lock()
|
||||||
|
|
||||||
def __getstate__(self):
|
def __getstate__(self):
|
||||||
"""RateLimiter is fully picklable, return full state."""
|
"""RateLimiter is fully picklable, return full state."""
|
||||||
return self.__dict__.copy()
|
state = self.__dict__.copy()
|
||||||
|
# Exclude unpickleable lock
|
||||||
|
if '_lock' in state:
|
||||||
|
del state['_lock']
|
||||||
|
return state
|
||||||
|
|
||||||
def __setstate__(self, state):
|
def __setstate__(self, state):
|
||||||
"""Restore RateLimiter state."""
|
"""Restore RateLimiter state."""
|
||||||
self.__dict__.update(state)
|
self.__dict__.update(state)
|
||||||
|
self._lock = threading.Lock()
|
||||||
|
|
||||||
def wait_if_needed(self) -> None:
|
def wait_if_needed(self) -> None:
|
||||||
"""Wait if necessary to respect rate limits."""
|
"""Wait if necessary to respect rate limits."""
|
||||||
current_time = time.time()
|
with self._lock:
|
||||||
time_since_last = current_time - self.last_request_time
|
current_time = time.time()
|
||||||
|
time_since_last = current_time - self.last_request_time
|
||||||
|
|
||||||
if time_since_last < self.min_interval:
|
if time_since_last < self.min_interval:
|
||||||
sleep_time = self.min_interval - time_since_last
|
sleep_time = self.min_interval - time_since_last
|
||||||
time.sleep(sleep_time)
|
time.sleep(sleep_time)
|
||||||
|
|
||||||
self.last_request_time = time.time()
|
self.last_request_time = time.time()
|
||||||
|
|
||||||
|
|
||||||
|
class ProviderCache:
|
||||||
|
"""Thread-safe global cache for provider queries."""
|
||||||
|
|
||||||
|
def __init__(self, provider_name: str, cache_expiry_hours: int = 12):
|
||||||
|
"""
|
||||||
|
Initialize provider-specific cache.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
provider_name: Name of the provider for cache directory
|
||||||
|
cache_expiry_hours: Cache expiry time in hours
|
||||||
|
"""
|
||||||
|
self.provider_name = provider_name
|
||||||
|
self.cache_expiry = cache_expiry_hours * 3600 # Convert to seconds
|
||||||
|
self.cache_dir = os.path.join('.cache', provider_name)
|
||||||
|
self._lock = threading.Lock()
|
||||||
|
|
||||||
|
# Ensure cache directory exists with thread-safe creation
|
||||||
|
os.makedirs(self.cache_dir, exist_ok=True)
|
||||||
|
|
||||||
|
def _generate_cache_key(self, method: str, url: str, params: Optional[Dict[str, Any]]) -> str:
|
||||||
|
"""Generate unique cache key for request."""
|
||||||
|
cache_data = f"{method}:{url}:{json.dumps(params or {}, sort_keys=True)}"
|
||||||
|
return hashlib.md5(cache_data.encode()).hexdigest() + ".json"
|
||||||
|
|
||||||
|
def get_cached_response(self, method: str, url: str, params: Optional[Dict[str, Any]]) -> Optional[requests.Response]:
|
||||||
|
"""
|
||||||
|
Retrieve cached response if available and not expired.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Cached Response object or None if cache miss/expired
|
||||||
|
"""
|
||||||
|
cache_key = self._generate_cache_key(method, url, params)
|
||||||
|
cache_path = os.path.join(self.cache_dir, cache_key)
|
||||||
|
|
||||||
|
with self._lock:
|
||||||
|
if not os.path.exists(cache_path):
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Check if cache is expired
|
||||||
|
cache_age = time.time() - os.path.getmtime(cache_path)
|
||||||
|
if cache_age >= self.cache_expiry:
|
||||||
|
try:
|
||||||
|
os.remove(cache_path)
|
||||||
|
except OSError:
|
||||||
|
pass # File might have been removed by another thread
|
||||||
|
return None
|
||||||
|
|
||||||
|
try:
|
||||||
|
with open(cache_path, 'r', encoding='utf-8') as f:
|
||||||
|
cached_data = json.load(f)
|
||||||
|
|
||||||
|
# Reconstruct Response object
|
||||||
|
response = requests.Response()
|
||||||
|
response.status_code = cached_data['status_code']
|
||||||
|
response._content = cached_data['content'].encode('utf-8')
|
||||||
|
response.headers.update(cached_data['headers'])
|
||||||
|
|
||||||
|
return response
|
||||||
|
|
||||||
|
except (json.JSONDecodeError, KeyError, IOError) as e:
|
||||||
|
# Cache file corrupted, remove it
|
||||||
|
try:
|
||||||
|
os.remove(cache_path)
|
||||||
|
except OSError:
|
||||||
|
pass
|
||||||
|
return None
|
||||||
|
|
||||||
|
def cache_response(self, method: str, url: str, params: Optional[Dict[str, Any]],
|
||||||
|
response: requests.Response) -> bool:
|
||||||
|
"""
|
||||||
|
Cache successful response to disk.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
True if cached successfully, False otherwise
|
||||||
|
"""
|
||||||
|
if response.status_code != 200:
|
||||||
|
return False
|
||||||
|
|
||||||
|
cache_key = self._generate_cache_key(method, url, params)
|
||||||
|
cache_path = os.path.join(self.cache_dir, cache_key)
|
||||||
|
|
||||||
|
with self._lock:
|
||||||
|
try:
|
||||||
|
cache_data = {
|
||||||
|
'status_code': response.status_code,
|
||||||
|
'content': response.text,
|
||||||
|
'headers': dict(response.headers),
|
||||||
|
'cached_at': datetime.now(timezone.utc).isoformat()
|
||||||
|
}
|
||||||
|
|
||||||
|
# Write to temporary file first, then rename for atomic operation
|
||||||
|
temp_path = cache_path + '.tmp'
|
||||||
|
with open(temp_path, 'w', encoding='utf-8') as f:
|
||||||
|
json.dump(cache_data, f)
|
||||||
|
|
||||||
|
# Atomic rename to prevent partial cache files
|
||||||
|
os.rename(temp_path, cache_path)
|
||||||
|
return True
|
||||||
|
|
||||||
|
except (IOError, OSError) as e:
|
||||||
|
# Clean up temp file if it exists
|
||||||
|
try:
|
||||||
|
if os.path.exists(temp_path):
|
||||||
|
os.remove(temp_path)
|
||||||
|
except OSError:
|
||||||
|
pass
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
class BaseProvider(ABC):
|
class BaseProvider(ABC):
|
||||||
"""
|
"""
|
||||||
Abstract base class for all DNSRecon data providers.
|
Abstract base class for all DNSRecon data providers.
|
||||||
Now supports session-specific configuration.
|
Now supports global provider-specific caching and session-specific configuration.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, name: str, rate_limit: int = 60, timeout: int = 30, session_config=None):
|
def __init__(self, name: str, rate_limit: int = 60, timeout: int = 30, session_config=None):
|
||||||
"""
|
"""
|
||||||
Initialize base provider with session-specific configuration.
|
Initialize base provider with global caching and session-specific configuration.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
name: Provider name for logging
|
name: Provider name for logging
|
||||||
@@ -78,20 +198,25 @@ class BaseProvider(ABC):
|
|||||||
self.logger = get_forensic_logger()
|
self.logger = get_forensic_logger()
|
||||||
self._stop_event = None
|
self._stop_event = None
|
||||||
|
|
||||||
|
# GLOBAL provider-specific caching (not session-based)
|
||||||
|
self.cache = ProviderCache(name, cache_expiry_hours=12)
|
||||||
|
|
||||||
# Statistics (per provider instance)
|
# Statistics (per provider instance)
|
||||||
self.total_requests = 0
|
self.total_requests = 0
|
||||||
self.successful_requests = 0
|
self.successful_requests = 0
|
||||||
self.failed_requests = 0
|
self.failed_requests = 0
|
||||||
self.total_relationships_found = 0
|
self.total_relationships_found = 0
|
||||||
|
self.cache_hits = 0
|
||||||
|
self.cache_misses = 0
|
||||||
|
|
||||||
|
print(f"Initialized {name} provider with global cache and session config (rate: {actual_rate_limit}/min)")
|
||||||
|
|
||||||
def __getstate__(self):
|
def __getstate__(self):
|
||||||
"""Prepare BaseProvider for pickling by excluding unpicklable objects."""
|
"""Prepare BaseProvider for pickling by excluding unpicklable objects."""
|
||||||
state = self.__dict__.copy()
|
state = self.__dict__.copy()
|
||||||
# Exclude the unpickleable '_local' attribute and stop event
|
# Exclude the unpickleable '_local' attribute and stop event
|
||||||
unpicklable_attrs = ['_local', '_stop_event']
|
state['_local'] = None
|
||||||
for attr in unpicklable_attrs:
|
state['_stop_event'] = None
|
||||||
if attr in state:
|
|
||||||
del state[attr]
|
|
||||||
return state
|
return state
|
||||||
|
|
||||||
def __setstate__(self, state):
|
def __setstate__(self, state):
|
||||||
@@ -106,7 +231,7 @@ class BaseProvider(ABC):
|
|||||||
if not hasattr(self._local, 'session'):
|
if not hasattr(self._local, 'session'):
|
||||||
self._local.session = requests.Session()
|
self._local.session = requests.Session()
|
||||||
self._local.session.headers.update({
|
self._local.session.headers.update({
|
||||||
'User-Agent': 'DNSRecon/1.0 (Passive Reconnaissance Tool)'
|
'User-Agent': 'DNSRecon/2.0 (Passive Reconnaissance Tool)'
|
||||||
})
|
})
|
||||||
return self._local.session
|
return self._local.session
|
||||||
|
|
||||||
@@ -164,79 +289,153 @@ class BaseProvider(ABC):
|
|||||||
def make_request(self, url: str, method: str = "GET",
|
def make_request(self, url: str, method: str = "GET",
|
||||||
params: Optional[Dict[str, Any]] = None,
|
params: Optional[Dict[str, Any]] = None,
|
||||||
headers: Optional[Dict[str, str]] = None,
|
headers: Optional[Dict[str, str]] = None,
|
||||||
target_indicator: str = "") -> Optional[requests.Response]:
|
target_indicator: str = "",
|
||||||
|
max_retries: int = 3) -> Optional[requests.Response]:
|
||||||
"""
|
"""
|
||||||
Make a rate-limited HTTP request.
|
Make a rate-limited HTTP request with global caching and aggressive stop signal handling.
|
||||||
"""
|
"""
|
||||||
|
# Check for cancellation before starting
|
||||||
if self._is_stop_requested():
|
if self._is_stop_requested():
|
||||||
print(f"Request cancelled before start: {url}")
|
print(f"Request cancelled before start: {url}")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
self.rate_limiter.wait_if_needed()
|
# Check global cache first
|
||||||
|
cached_response = self.cache.get_cached_response(method, url, params)
|
||||||
|
if cached_response is not None:
|
||||||
|
print(f"Cache hit for {self.name}: {url}")
|
||||||
|
self.cache_hits += 1
|
||||||
|
return cached_response
|
||||||
|
|
||||||
|
self.cache_misses += 1
|
||||||
|
|
||||||
start_time = time.time()
|
# Determine effective max_retries based on stop signal
|
||||||
response = None
|
effective_max_retries = 0 if self._is_stop_requested() else max_retries
|
||||||
error = None
|
last_exception = None
|
||||||
|
|
||||||
try:
|
for attempt in range(effective_max_retries + 1):
|
||||||
self.total_requests += 1
|
# Check for cancellation before each attempt
|
||||||
|
if self._is_stop_requested():
|
||||||
|
print(f"Request cancelled during attempt {attempt + 1}: {url}")
|
||||||
|
return None
|
||||||
|
|
||||||
request_headers = dict(self.session.headers).copy()
|
# Apply rate limiting with cancellation awareness
|
||||||
if headers:
|
if not self._wait_with_cancellation_check():
|
||||||
request_headers.update(headers)
|
print(f"Request cancelled during rate limiting: {url}")
|
||||||
|
return None
|
||||||
|
|
||||||
print(f"Making {method} request to: {url}")
|
# Final check before making HTTP request
|
||||||
|
if self._is_stop_requested():
|
||||||
|
print(f"Request cancelled before HTTP call: {url}")
|
||||||
|
return None
|
||||||
|
|
||||||
if method.upper() == "GET":
|
start_time = time.time()
|
||||||
response = self.session.get(
|
response = None
|
||||||
url,
|
error = None
|
||||||
params=params,
|
|
||||||
headers=request_headers,
|
try:
|
||||||
timeout=self.timeout
|
self.total_requests += 1
|
||||||
|
|
||||||
|
# Prepare request
|
||||||
|
request_headers = self.session.headers.copy()
|
||||||
|
if headers:
|
||||||
|
request_headers.update(headers)
|
||||||
|
|
||||||
|
print(f"Making {method} request to: {url} (attempt {attempt + 1})")
|
||||||
|
|
||||||
|
# Use shorter timeout if termination is requested
|
||||||
|
request_timeout = 2 if self._is_stop_requested() else self.timeout
|
||||||
|
|
||||||
|
# Make request
|
||||||
|
if method.upper() == "GET":
|
||||||
|
response = self.session.get(
|
||||||
|
url,
|
||||||
|
params=params,
|
||||||
|
headers=request_headers,
|
||||||
|
timeout=request_timeout
|
||||||
|
)
|
||||||
|
elif method.upper() == "POST":
|
||||||
|
response = self.session.post(
|
||||||
|
url,
|
||||||
|
json=params,
|
||||||
|
headers=request_headers,
|
||||||
|
timeout=request_timeout
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
raise ValueError(f"Unsupported HTTP method: {method}")
|
||||||
|
|
||||||
|
print(f"Response status: {response.status_code}")
|
||||||
|
response.raise_for_status()
|
||||||
|
self.successful_requests += 1
|
||||||
|
|
||||||
|
# Success - log, cache, and return
|
||||||
|
duration_ms = (time.time() - start_time) * 1000
|
||||||
|
self.logger.log_api_request(
|
||||||
|
provider=self.name,
|
||||||
|
url=url,
|
||||||
|
method=method.upper(),
|
||||||
|
status_code=response.status_code,
|
||||||
|
response_size=len(response.content),
|
||||||
|
duration_ms=duration_ms,
|
||||||
|
error=None,
|
||||||
|
target_indicator=target_indicator
|
||||||
)
|
)
|
||||||
elif method.upper() == "POST":
|
|
||||||
response = self.session.post(
|
# Cache the successful response globally
|
||||||
url,
|
self.cache.cache_response(method, url, params, response)
|
||||||
json=params,
|
return response
|
||||||
headers=request_headers,
|
|
||||||
timeout=self.timeout
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
raise ValueError(f"Unsupported HTTP method: {method}")
|
|
||||||
|
|
||||||
print(f"Response status: {response.status_code}")
|
except requests.exceptions.RequestException as e:
|
||||||
response.raise_for_status()
|
error = str(e)
|
||||||
self.successful_requests += 1
|
self.failed_requests += 1
|
||||||
|
print(f"Request failed (attempt {attempt + 1}): {error}")
|
||||||
duration_ms = (time.time() - start_time) * 1000
|
last_exception = e
|
||||||
self.logger.log_api_request(
|
|
||||||
provider=self.name,
|
# Immediately abort retries if stop requested
|
||||||
url=url,
|
if self._is_stop_requested():
|
||||||
method=method.upper(),
|
print(f"Stop requested - aborting retries for: {url}")
|
||||||
status_code=response.status_code,
|
break
|
||||||
response_size=len(response.content),
|
|
||||||
duration_ms=duration_ms,
|
# Check if we should retry
|
||||||
error=None,
|
if attempt < effective_max_retries and self._should_retry(e):
|
||||||
target_indicator=target_indicator
|
# Exponential backoff with jitter for 429 errors
|
||||||
)
|
if isinstance(e, requests.exceptions.HTTPError) and e.response and e.response.status_code == 429:
|
||||||
|
backoff_time = min(60, 10 * (2 ** attempt))
|
||||||
return response
|
print(f"Rate limit hit. Retrying in {backoff_time} seconds...")
|
||||||
|
else:
|
||||||
|
backoff_time = min(2.0, (2 ** attempt) * 0.5)
|
||||||
|
print(f"Retrying in {backoff_time} seconds...")
|
||||||
|
|
||||||
|
if not self._sleep_with_cancellation_check(backoff_time):
|
||||||
|
print(f"Stop requested during backoff - aborting: {url}")
|
||||||
|
return None
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
break
|
||||||
|
|
||||||
except requests.exceptions.RequestException as e:
|
except Exception as e:
|
||||||
error = str(e)
|
error = f"Unexpected error: {str(e)}"
|
||||||
self.failed_requests += 1
|
self.failed_requests += 1
|
||||||
duration_ms = (time.time() - start_time) * 1000
|
print(f"Unexpected error: {error}")
|
||||||
self.logger.log_api_request(
|
last_exception = e
|
||||||
provider=self.name,
|
break
|
||||||
url=url,
|
|
||||||
method=method.upper(),
|
# All attempts failed - log and return None
|
||||||
status_code=response.status_code if response else None,
|
duration_ms = (time.time() - start_time) * 1000
|
||||||
response_size=len(response.content) if response else None,
|
self.logger.log_api_request(
|
||||||
duration_ms=duration_ms,
|
provider=self.name,
|
||||||
error=error,
|
url=url,
|
||||||
target_indicator=target_indicator
|
method=method.upper(),
|
||||||
)
|
status_code=response.status_code if response else None,
|
||||||
raise e
|
response_size=len(response.content) if response else None,
|
||||||
|
duration_ms=duration_ms,
|
||||||
|
error=error,
|
||||||
|
target_indicator=target_indicator
|
||||||
|
)
|
||||||
|
|
||||||
|
if error and last_exception:
|
||||||
|
raise last_exception
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
def _is_stop_requested(self) -> bool:
|
def _is_stop_requested(self) -> bool:
|
||||||
"""
|
"""
|
||||||
@@ -246,6 +445,43 @@ class BaseProvider(ABC):
|
|||||||
return True
|
return True
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
def _wait_with_cancellation_check(self) -> bool:
|
||||||
|
"""
|
||||||
|
Wait for rate limiting while aggressively checking for cancellation.
|
||||||
|
Returns False if cancelled during wait.
|
||||||
|
"""
|
||||||
|
current_time = time.time()
|
||||||
|
time_since_last = current_time - self.rate_limiter.last_request_time
|
||||||
|
|
||||||
|
if time_since_last < self.rate_limiter.min_interval:
|
||||||
|
sleep_time = self.rate_limiter.min_interval - time_since_last
|
||||||
|
if not self._sleep_with_cancellation_check(sleep_time):
|
||||||
|
return False
|
||||||
|
|
||||||
|
self.rate_limiter.last_request_time = time.time()
|
||||||
|
return True
|
||||||
|
|
||||||
|
def _sleep_with_cancellation_check(self, sleep_time: float) -> bool:
|
||||||
|
"""
|
||||||
|
Sleep for the specified time while aggressively checking for cancellation.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
sleep_time: Time to sleep in seconds
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
bool: True if sleep completed, False if cancelled
|
||||||
|
"""
|
||||||
|
sleep_start = time.time()
|
||||||
|
check_interval = 0.05 # Check every 50ms for aggressive responsiveness
|
||||||
|
|
||||||
|
while time.time() - sleep_start < sleep_time:
|
||||||
|
if self._is_stop_requested():
|
||||||
|
return False
|
||||||
|
remaining_time = sleep_time - (time.time() - sleep_start)
|
||||||
|
time.sleep(min(check_interval, remaining_time))
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
def set_stop_event(self, stop_event: threading.Event) -> None:
|
def set_stop_event(self, stop_event: threading.Event) -> None:
|
||||||
"""
|
"""
|
||||||
Set the stop event for this provider to enable cancellation.
|
Set the stop event for this provider to enable cancellation.
|
||||||
@@ -255,6 +491,28 @@ class BaseProvider(ABC):
|
|||||||
"""
|
"""
|
||||||
self._stop_event = stop_event
|
self._stop_event = stop_event
|
||||||
|
|
||||||
|
def _should_retry(self, exception: requests.exceptions.RequestException) -> bool:
|
||||||
|
"""
|
||||||
|
Determine if a request should be retried based on the exception.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
exception: The request exception that occurred
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
True if the request should be retried
|
||||||
|
"""
|
||||||
|
# Retry on connection errors and timeouts
|
||||||
|
if isinstance(exception, (requests.exceptions.ConnectionError,
|
||||||
|
requests.exceptions.Timeout)):
|
||||||
|
return True
|
||||||
|
|
||||||
|
if isinstance(exception, requests.exceptions.HTTPError):
|
||||||
|
if hasattr(exception, 'response') and exception.response:
|
||||||
|
# Retry on server errors (5xx) AND on rate-limiting errors (429)
|
||||||
|
return exception.response.status_code >= 500 or exception.response.status_code == 429
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
def log_relationship_discovery(self, source_node: str, target_node: str,
|
def log_relationship_discovery(self, source_node: str, target_node: str,
|
||||||
relationship_type: str,
|
relationship_type: str,
|
||||||
confidence_score: float,
|
confidence_score: float,
|
||||||
@@ -285,7 +543,7 @@ class BaseProvider(ABC):
|
|||||||
|
|
||||||
def get_statistics(self) -> Dict[str, Any]:
|
def get_statistics(self) -> Dict[str, Any]:
|
||||||
"""
|
"""
|
||||||
Get provider statistics.
|
Get provider statistics including cache performance.
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
Dictionary containing provider performance metrics
|
Dictionary containing provider performance metrics
|
||||||
@@ -297,5 +555,8 @@ class BaseProvider(ABC):
|
|||||||
'failed_requests': self.failed_requests,
|
'failed_requests': self.failed_requests,
|
||||||
'success_rate': (self.successful_requests / self.total_requests * 100) if self.total_requests > 0 else 0,
|
'success_rate': (self.successful_requests / self.total_requests * 100) if self.total_requests > 0 else 0,
|
||||||
'relationships_found': self.total_relationships_found,
|
'relationships_found': self.total_relationships_found,
|
||||||
'rate_limit': self.rate_limiter.requests_per_minute
|
'rate_limit': self.rate_limiter.requests_per_minute,
|
||||||
|
'cache_hits': self.cache_hits,
|
||||||
|
'cache_misses': self.cache_misses,
|
||||||
|
'cache_hit_rate': (self.cache_hits / (self.cache_hits + self.cache_misses) * 100) if (self.cache_hits + self.cache_misses) > 0 else 0
|
||||||
}
|
}
|
||||||
@@ -1,9 +1,11 @@
|
|||||||
# dnsrecon/providers/crtsh_provider.py
|
"""
|
||||||
|
Certificate Transparency provider using crt.sh.
|
||||||
|
Discovers domain relationships through certificate SAN analysis with comprehensive certificate tracking.
|
||||||
|
Stores certificates as metadata on domain nodes rather than creating certificate nodes.
|
||||||
|
"""
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import re
|
import re
|
||||||
import os
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import List, Dict, Any, Tuple, Set
|
from typing import List, Dict, Any, Tuple, Set
|
||||||
from urllib.parse import quote
|
from urllib.parse import quote
|
||||||
from datetime import datetime, timezone
|
from datetime import datetime, timezone
|
||||||
@@ -16,10 +18,10 @@ from utils.helpers import _is_valid_domain
|
|||||||
class CrtShProvider(BaseProvider):
|
class CrtShProvider(BaseProvider):
|
||||||
"""
|
"""
|
||||||
Provider for querying crt.sh certificate transparency database.
|
Provider for querying crt.sh certificate transparency database.
|
||||||
Now uses session-specific configuration and caching with accumulative behavior.
|
Now uses session-specific configuration and caching.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, name=None, session_config=None):
|
def __init__(self, session_config=None):
|
||||||
"""Initialize CrtSh provider with session-specific configuration."""
|
"""Initialize CrtSh provider with session-specific configuration."""
|
||||||
super().__init__(
|
super().__init__(
|
||||||
name="crtsh",
|
name="crtsh",
|
||||||
@@ -29,10 +31,6 @@ class CrtShProvider(BaseProvider):
|
|||||||
)
|
)
|
||||||
self.base_url = "https://crt.sh/"
|
self.base_url = "https://crt.sh/"
|
||||||
self._stop_event = None
|
self._stop_event = None
|
||||||
|
|
||||||
# Initialize cache directory
|
|
||||||
self.cache_dir = Path('cache') / 'crtsh'
|
|
||||||
self.cache_dir.mkdir(parents=True, exist_ok=True)
|
|
||||||
|
|
||||||
def get_name(self) -> str:
|
def get_name(self) -> str:
|
||||||
"""Return the provider name."""
|
"""Return the provider name."""
|
||||||
@@ -58,158 +56,6 @@ class CrtShProvider(BaseProvider):
|
|||||||
"""
|
"""
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def _get_cache_file_path(self, domain: str) -> Path:
|
|
||||||
"""Generate cache file path for a domain."""
|
|
||||||
# Sanitize domain for filename safety
|
|
||||||
safe_domain = domain.replace('.', '_').replace('/', '_').replace('\\', '_')
|
|
||||||
return self.cache_dir / f"{safe_domain}.json"
|
|
||||||
|
|
||||||
def _get_cache_status(self, cache_file_path: Path) -> str:
|
|
||||||
"""
|
|
||||||
Check cache status for a domain.
|
|
||||||
Returns: 'not_found', 'fresh', or 'stale'
|
|
||||||
"""
|
|
||||||
if not cache_file_path.exists():
|
|
||||||
return "not_found"
|
|
||||||
|
|
||||||
try:
|
|
||||||
with open(cache_file_path, 'r') as f:
|
|
||||||
cache_data = json.load(f)
|
|
||||||
|
|
||||||
last_query_str = cache_data.get("last_upstream_query")
|
|
||||||
if not last_query_str:
|
|
||||||
return "stale" # Invalid cache format
|
|
||||||
|
|
||||||
last_query = datetime.fromisoformat(last_query_str.replace('Z', '+00:00'))
|
|
||||||
hours_since_query = (datetime.now(timezone.utc) - last_query).total_seconds() / 3600
|
|
||||||
|
|
||||||
cache_timeout = self.config.cache_timeout_hours
|
|
||||||
if hours_since_query < cache_timeout:
|
|
||||||
return "fresh"
|
|
||||||
else:
|
|
||||||
return "stale"
|
|
||||||
|
|
||||||
except (json.JSONDecodeError, ValueError, KeyError) as e:
|
|
||||||
self.logger.logger.warning(f"Invalid cache file format for {cache_file_path}: {e}")
|
|
||||||
return "stale"
|
|
||||||
|
|
||||||
def _load_cached_certificates(self, cache_file_path: Path) -> List[Dict[str, Any]]:
|
|
||||||
"""Load certificates from cache file."""
|
|
||||||
try:
|
|
||||||
with open(cache_file_path, 'r') as f:
|
|
||||||
cache_data = json.load(f)
|
|
||||||
return cache_data.get('certificates', [])
|
|
||||||
except (json.JSONDecodeError, FileNotFoundError, KeyError) as e:
|
|
||||||
self.logger.logger.error(f"Failed to load cached certificates from {cache_file_path}: {e}")
|
|
||||||
return []
|
|
||||||
|
|
||||||
def _query_crtsh_api(self, domain: str) -> List[Dict[str, Any]]:
|
|
||||||
"""
|
|
||||||
Query crt.sh API for raw certificate data.
|
|
||||||
Raises exceptions for network errors to allow core logic to retry.
|
|
||||||
"""
|
|
||||||
url = f"{self.base_url}?q={quote(domain)}&output=json"
|
|
||||||
response = self.make_request(url, target_indicator=domain)
|
|
||||||
|
|
||||||
if not response or response.status_code != 200:
|
|
||||||
# This could be a temporary error - raise exception so core can retry
|
|
||||||
raise requests.exceptions.RequestException(f"crt.sh API returned status {response.status_code if response else 'None'}")
|
|
||||||
|
|
||||||
certificates = response.json()
|
|
||||||
if not certificates:
|
|
||||||
return []
|
|
||||||
|
|
||||||
return certificates
|
|
||||||
|
|
||||||
def _create_cache_file(self, cache_file_path: Path, domain: str, certificates: List[Dict[str, Any]]) -> None:
|
|
||||||
"""Create new cache file with certificates."""
|
|
||||||
try:
|
|
||||||
cache_data = {
|
|
||||||
"domain": domain,
|
|
||||||
"first_cached": datetime.now(timezone.utc).isoformat(),
|
|
||||||
"last_upstream_query": datetime.now(timezone.utc).isoformat(),
|
|
||||||
"upstream_query_count": 1,
|
|
||||||
"certificates": certificates
|
|
||||||
}
|
|
||||||
|
|
||||||
cache_file_path.parent.mkdir(parents=True, exist_ok=True)
|
|
||||||
with open(cache_file_path, 'w') as f:
|
|
||||||
json.dump(cache_data, f, separators=(',', ':'))
|
|
||||||
|
|
||||||
self.logger.logger.info(f"Created cache file for {domain} with {len(certificates)} certificates")
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.logger.warning(f"Failed to create cache file for {domain}: {e}")
|
|
||||||
|
|
||||||
def _append_to_cache(self, cache_file_path: Path, new_certificates: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
|
|
||||||
"""Append new certificates to existing cache and return all certificates."""
|
|
||||||
try:
|
|
||||||
# Load existing cache
|
|
||||||
with open(cache_file_path, 'r') as f:
|
|
||||||
cache_data = json.load(f)
|
|
||||||
|
|
||||||
# Track existing certificate IDs to avoid duplicates
|
|
||||||
existing_ids = {cert.get('id') for cert in cache_data.get('certificates', [])}
|
|
||||||
|
|
||||||
# Add only new certificates
|
|
||||||
added_count = 0
|
|
||||||
for cert in new_certificates:
|
|
||||||
cert_id = cert.get('id')
|
|
||||||
if cert_id and cert_id not in existing_ids:
|
|
||||||
cache_data['certificates'].append(cert)
|
|
||||||
existing_ids.add(cert_id)
|
|
||||||
added_count += 1
|
|
||||||
|
|
||||||
# Update metadata
|
|
||||||
cache_data['last_upstream_query'] = datetime.now(timezone.utc).isoformat()
|
|
||||||
cache_data['upstream_query_count'] = cache_data.get('upstream_query_count', 0) + 1
|
|
||||||
|
|
||||||
# Write updated cache
|
|
||||||
with open(cache_file_path, 'w') as f:
|
|
||||||
json.dump(cache_data, f, separators=(',', ':'))
|
|
||||||
|
|
||||||
total_certs = len(cache_data['certificates'])
|
|
||||||
self.logger.logger.info(f"Appended {added_count} new certificates to cache. Total: {total_certs}")
|
|
||||||
|
|
||||||
return cache_data['certificates']
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.logger.warning(f"Failed to append to cache: {e}")
|
|
||||||
return new_certificates # Fallback to new certificates only
|
|
||||||
|
|
||||||
def _parse_issuer_organization(self, issuer_dn: str) -> str:
|
|
||||||
"""
|
|
||||||
Parse the issuer Distinguished Name to extract just the organization name.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
issuer_dn: Full issuer DN string (e.g., "C=US, O=Let's Encrypt, CN=R11")
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Organization name (e.g., "Let's Encrypt") or original string if parsing fails
|
|
||||||
"""
|
|
||||||
if not issuer_dn:
|
|
||||||
return issuer_dn
|
|
||||||
|
|
||||||
try:
|
|
||||||
# Split by comma and look for O= component
|
|
||||||
components = [comp.strip() for comp in issuer_dn.split(',')]
|
|
||||||
|
|
||||||
for component in components:
|
|
||||||
if component.startswith('O='):
|
|
||||||
# Extract the value after O=
|
|
||||||
org_name = component[2:].strip()
|
|
||||||
# Remove quotes if present
|
|
||||||
if org_name.startswith('"') and org_name.endswith('"'):
|
|
||||||
org_name = org_name[1:-1]
|
|
||||||
return org_name
|
|
||||||
|
|
||||||
# If no O= component found, return the original string
|
|
||||||
return issuer_dn
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.logger.debug(f"Failed to parse issuer DN '{issuer_dn}': {e}")
|
|
||||||
return issuer_dn
|
|
||||||
|
|
||||||
def _parse_certificate_date(self, date_string: str) -> datetime:
|
def _parse_certificate_date(self, date_string: str) -> datetime:
|
||||||
"""
|
"""
|
||||||
Parse certificate date from crt.sh format.
|
Parse certificate date from crt.sh format.
|
||||||
@@ -287,15 +133,10 @@ class CrtShProvider(BaseProvider):
|
|||||||
Returns:
|
Returns:
|
||||||
Comprehensive certificate metadata dictionary
|
Comprehensive certificate metadata dictionary
|
||||||
"""
|
"""
|
||||||
# Parse the issuer name to get just the organization
|
|
||||||
raw_issuer_name = cert_data.get('issuer_name', '')
|
|
||||||
parsed_issuer_name = self._parse_issuer_organization(raw_issuer_name)
|
|
||||||
|
|
||||||
metadata = {
|
metadata = {
|
||||||
'certificate_id': cert_data.get('id'),
|
'certificate_id': cert_data.get('id'),
|
||||||
'serial_number': cert_data.get('serial_number'),
|
'serial_number': cert_data.get('serial_number'),
|
||||||
'issuer_name': parsed_issuer_name, # Use parsed organization name
|
'issuer_name': cert_data.get('issuer_name'),
|
||||||
#'issuer_name_full': raw_issuer_name, # deliberately left out, because its not useful in most cases
|
|
||||||
'issuer_ca_id': cert_data.get('issuer_ca_id'),
|
'issuer_ca_id': cert_data.get('issuer_ca_id'),
|
||||||
'common_name': cert_data.get('common_name'),
|
'common_name': cert_data.get('common_name'),
|
||||||
'not_before': cert_data.get('not_before'),
|
'not_before': cert_data.get('not_before'),
|
||||||
@@ -326,8 +167,7 @@ class CrtShProvider(BaseProvider):
|
|||||||
|
|
||||||
def query_domain(self, domain: str) -> List[Tuple[str, str, str, float, Dict[str, Any]]]:
|
def query_domain(self, domain: str) -> List[Tuple[str, str, str, float, Dict[str, Any]]]:
|
||||||
"""
|
"""
|
||||||
Query crt.sh for certificates containing the domain with caching support.
|
Query crt.sh for certificates containing the domain.
|
||||||
Properly raises exceptions for network errors to allow core logic retries.
|
|
||||||
"""
|
"""
|
||||||
if not _is_valid_domain(domain):
|
if not _is_valid_domain(domain):
|
||||||
return []
|
return []
|
||||||
@@ -337,167 +177,131 @@ class CrtShProvider(BaseProvider):
|
|||||||
print(f"CrtSh query cancelled before start for domain: {domain}")
|
print(f"CrtSh query cancelled before start for domain: {domain}")
|
||||||
return []
|
return []
|
||||||
|
|
||||||
# === CACHING LOGIC ===
|
|
||||||
cache_file = self._get_cache_file_path(domain)
|
|
||||||
cache_status = self._get_cache_status(cache_file)
|
|
||||||
|
|
||||||
certificates = []
|
|
||||||
|
|
||||||
try:
|
|
||||||
if cache_status == "fresh":
|
|
||||||
# Use cached data
|
|
||||||
certificates = self._load_cached_certificates(cache_file)
|
|
||||||
self.logger.logger.info(f"Using cached data for {domain} ({len(certificates)} certificates)")
|
|
||||||
|
|
||||||
elif cache_status == "not_found":
|
|
||||||
# Fresh query, create new cache
|
|
||||||
certificates = self._query_crtsh_api(domain)
|
|
||||||
if certificates: # Only cache if we got results
|
|
||||||
self._create_cache_file(cache_file, domain, certificates)
|
|
||||||
self.logger.logger.info(f"Cached fresh data for {domain} ({len(certificates)} certificates)")
|
|
||||||
else:
|
|
||||||
self.logger.logger.info(f"No certificates found for {domain}, not caching")
|
|
||||||
|
|
||||||
elif cache_status == "stale":
|
|
||||||
# Append query, update existing cache
|
|
||||||
try:
|
|
||||||
new_certificates = self._query_crtsh_api(domain)
|
|
||||||
if new_certificates:
|
|
||||||
certificates = self._append_to_cache(cache_file, new_certificates)
|
|
||||||
self.logger.logger.info(f"Refreshed and appended cache for {domain}")
|
|
||||||
else:
|
|
||||||
# Use existing cache if API returns no results
|
|
||||||
certificates = self._load_cached_certificates(cache_file)
|
|
||||||
self.logger.logger.info(f"API returned no new results, using existing cache for {domain}")
|
|
||||||
except requests.exceptions.RequestException:
|
|
||||||
# If API call fails for stale cache, use cached data and re-raise for retry logic
|
|
||||||
certificates = self._load_cached_certificates(cache_file)
|
|
||||||
if certificates:
|
|
||||||
self.logger.logger.warning(f"API call failed for {domain}, using stale cache data ({len(certificates)} certificates)")
|
|
||||||
# Don't re-raise here, just use cached data
|
|
||||||
else:
|
|
||||||
# No cached data and API failed - re-raise for retry
|
|
||||||
raise
|
|
||||||
|
|
||||||
except requests.exceptions.RequestException as e:
|
|
||||||
# Network/API errors should be re-raised so core logic can retry
|
|
||||||
self.logger.logger.error(f"API query failed for {domain}: {e}")
|
|
||||||
raise e
|
|
||||||
except json.JSONDecodeError as e:
|
|
||||||
# JSON parsing errors should also be raised for retry
|
|
||||||
self.logger.logger.error(f"Failed to parse JSON response from crt.sh for {domain}: {e}")
|
|
||||||
raise e
|
|
||||||
|
|
||||||
# Check for cancellation after cache operations
|
|
||||||
if self._stop_event and self._stop_event.is_set():
|
|
||||||
print(f"CrtSh query cancelled after cache operations for domain: {domain}")
|
|
||||||
return []
|
|
||||||
|
|
||||||
if not certificates:
|
|
||||||
return []
|
|
||||||
|
|
||||||
return self._process_certificates_to_relationships(domain, certificates)
|
|
||||||
|
|
||||||
def _process_certificates_to_relationships(self, domain: str, certificates: List[Dict[str, Any]]) -> List[Tuple[str, str, str, float, Dict[str, Any]]]:
|
|
||||||
"""
|
|
||||||
Process certificates to relationships using existing logic.
|
|
||||||
This method contains the original processing logic from query_domain.
|
|
||||||
"""
|
|
||||||
relationships = []
|
relationships = []
|
||||||
|
|
||||||
# Check for cancellation before processing
|
try:
|
||||||
if self._stop_event and self._stop_event.is_set():
|
# Query crt.sh for certificates
|
||||||
print(f"CrtSh processing cancelled before processing for domain: {domain}")
|
url = f"{self.base_url}?q={quote(domain)}&output=json"
|
||||||
return []
|
response = self.make_request(url, target_indicator=domain, max_retries=3)
|
||||||
|
|
||||||
# Aggregate certificate data by domain
|
|
||||||
domain_certificates = {}
|
|
||||||
all_discovered_domains = set()
|
|
||||||
|
|
||||||
# Process certificates with cancellation checking
|
|
||||||
for i, cert_data in enumerate(certificates):
|
|
||||||
# Check for cancellation every 5 certificates for faster response
|
|
||||||
if i % 5 == 0 and self._stop_event and self._stop_event.is_set():
|
|
||||||
print(f"CrtSh processing cancelled at certificate {i} for domain: {domain}")
|
|
||||||
break
|
|
||||||
|
|
||||||
cert_metadata = self._extract_certificate_metadata(cert_data)
|
|
||||||
cert_domains = self._extract_domains_from_certificate(cert_data)
|
|
||||||
|
|
||||||
# Add all domains from this certificate to our tracking
|
if not response or response.status_code != 200:
|
||||||
all_discovered_domains.update(cert_domains)
|
return []
|
||||||
for cert_domain in cert_domains:
|
|
||||||
if not _is_valid_domain(cert_domain):
|
# Check for cancellation after request
|
||||||
|
if self._stop_event and self._stop_event.is_set():
|
||||||
|
print(f"CrtSh query cancelled after request for domain: {domain}")
|
||||||
|
return []
|
||||||
|
|
||||||
|
certificates = response.json()
|
||||||
|
|
||||||
|
if not certificates:
|
||||||
|
return []
|
||||||
|
|
||||||
|
# Check for cancellation before processing
|
||||||
|
if self._stop_event and self._stop_event.is_set():
|
||||||
|
print(f"CrtSh query cancelled before processing for domain: {domain}")
|
||||||
|
return []
|
||||||
|
|
||||||
|
# Aggregate certificate data by domain
|
||||||
|
domain_certificates = {}
|
||||||
|
all_discovered_domains = set()
|
||||||
|
|
||||||
|
# Process certificates with cancellation checking
|
||||||
|
for i, cert_data in enumerate(certificates):
|
||||||
|
# Check for cancellation every 5 certificates instead of 10 for faster response
|
||||||
|
if i % 5 == 0 and self._stop_event and self._stop_event.is_set():
|
||||||
|
print(f"CrtSh processing cancelled at certificate {i} for domain: {domain}")
|
||||||
|
break
|
||||||
|
|
||||||
|
cert_metadata = self._extract_certificate_metadata(cert_data)
|
||||||
|
cert_domains = self._extract_domains_from_certificate(cert_data)
|
||||||
|
|
||||||
|
# Add all domains from this certificate to our tracking
|
||||||
|
for cert_domain in cert_domains:
|
||||||
|
# Additional stop check during domain processing
|
||||||
|
if i % 20 == 0 and self._stop_event and self._stop_event.is_set():
|
||||||
|
print(f"CrtSh domain processing cancelled for domain: {domain}")
|
||||||
|
break
|
||||||
|
|
||||||
|
if not _is_valid_domain(cert_domain):
|
||||||
|
continue
|
||||||
|
|
||||||
|
all_discovered_domains.add(cert_domain)
|
||||||
|
|
||||||
|
# Initialize domain certificate list if needed
|
||||||
|
if cert_domain not in domain_certificates:
|
||||||
|
domain_certificates[cert_domain] = []
|
||||||
|
|
||||||
|
# Add this certificate to the domain's certificate list
|
||||||
|
domain_certificates[cert_domain].append(cert_metadata)
|
||||||
|
|
||||||
|
# Final cancellation check before creating relationships
|
||||||
|
if self._stop_event and self._stop_event.is_set():
|
||||||
|
print(f"CrtSh query cancelled before relationship creation for domain: {domain}")
|
||||||
|
return []
|
||||||
|
|
||||||
|
# Create relationships from query domain to ALL discovered domains with stop checking
|
||||||
|
for i, discovered_domain in enumerate(all_discovered_domains):
|
||||||
|
if discovered_domain == domain:
|
||||||
|
continue # Skip self-relationships
|
||||||
|
|
||||||
|
# Check for cancellation every 10 relationships
|
||||||
|
if i % 10 == 0 and self._stop_event and self._stop_event.is_set():
|
||||||
|
print(f"CrtSh relationship creation cancelled for domain: {domain}")
|
||||||
|
break
|
||||||
|
|
||||||
|
if not _is_valid_domain(discovered_domain):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# Initialize domain certificate list if needed
|
# Get certificates for both domains
|
||||||
if cert_domain not in domain_certificates:
|
query_domain_certs = domain_certificates.get(domain, [])
|
||||||
domain_certificates[cert_domain] = []
|
discovered_domain_certs = domain_certificates.get(discovered_domain, [])
|
||||||
|
|
||||||
# Add this certificate to the domain's certificate list
|
# Find shared certificates (for metadata purposes)
|
||||||
domain_certificates[cert_domain].append(cert_metadata)
|
shared_certificates = self._find_shared_certificates(query_domain_certs, discovered_domain_certs)
|
||||||
|
|
||||||
# Final cancellation check before creating relationships
|
# Calculate confidence based on relationship type and shared certificates
|
||||||
if self._stop_event and self._stop_event.is_set():
|
confidence = self._calculate_domain_relationship_confidence(
|
||||||
print(f"CrtSh query cancelled before relationship creation for domain: {domain}")
|
domain, discovered_domain, shared_certificates, all_discovered_domains
|
||||||
return []
|
)
|
||||||
|
|
||||||
# Create relationships from query domain to ALL discovered domains with stop checking
|
# Create comprehensive raw data for the relationship
|
||||||
for i, discovered_domain in enumerate(all_discovered_domains):
|
relationship_raw_data = {
|
||||||
if discovered_domain == domain:
|
'relationship_type': 'certificate_discovery',
|
||||||
continue # Skip self-relationships
|
'shared_certificates': shared_certificates,
|
||||||
|
'total_shared_certs': len(shared_certificates),
|
||||||
# Check for cancellation every 10 relationships
|
'discovery_context': self._determine_relationship_context(discovered_domain, domain),
|
||||||
if i % 10 == 0 and self._stop_event and self._stop_event.is_set():
|
'domain_certificates': {
|
||||||
print(f"CrtSh relationship creation cancelled for domain: {domain}")
|
domain: self._summarize_certificates(query_domain_certs),
|
||||||
break
|
discovered_domain: self._summarize_certificates(discovered_domain_certs)
|
||||||
|
}
|
||||||
if not _is_valid_domain(discovered_domain):
|
|
||||||
continue
|
|
||||||
|
|
||||||
# Get certificates for both domains
|
|
||||||
query_domain_certs = domain_certificates.get(domain, [])
|
|
||||||
discovered_domain_certs = domain_certificates.get(discovered_domain, [])
|
|
||||||
|
|
||||||
# Find shared certificates (for metadata purposes)
|
|
||||||
shared_certificates = self._find_shared_certificates(query_domain_certs, discovered_domain_certs)
|
|
||||||
|
|
||||||
# Calculate confidence based on relationship type and shared certificates
|
|
||||||
confidence = self._calculate_domain_relationship_confidence(
|
|
||||||
domain, discovered_domain, shared_certificates, all_discovered_domains
|
|
||||||
)
|
|
||||||
|
|
||||||
# Create comprehensive raw data for the relationship
|
|
||||||
relationship_raw_data = {
|
|
||||||
'relationship_type': 'certificate_discovery',
|
|
||||||
'shared_certificates': shared_certificates,
|
|
||||||
'total_shared_certs': len(shared_certificates),
|
|
||||||
'discovery_context': self._determine_relationship_context(discovered_domain, domain),
|
|
||||||
'domain_certificates': {
|
|
||||||
domain: self._summarize_certificates(query_domain_certs),
|
|
||||||
discovered_domain: self._summarize_certificates(discovered_domain_certs)
|
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
# Create domain -> domain relationship
|
||||||
# Create domain -> domain relationship
|
relationships.append((
|
||||||
relationships.append((
|
domain,
|
||||||
domain,
|
discovered_domain,
|
||||||
discovered_domain,
|
'san_certificate',
|
||||||
'san_certificate',
|
confidence,
|
||||||
confidence,
|
relationship_raw_data
|
||||||
relationship_raw_data
|
))
|
||||||
))
|
|
||||||
|
# Log the relationship discovery
|
||||||
# Log the relationship discovery
|
self.log_relationship_discovery(
|
||||||
self.log_relationship_discovery(
|
source_node=domain,
|
||||||
source_node=domain,
|
target_node=discovered_domain,
|
||||||
target_node=discovered_domain,
|
relationship_type='san_certificate',
|
||||||
relationship_type='san_certificate',
|
confidence_score=confidence,
|
||||||
confidence_score=confidence,
|
raw_data=relationship_raw_data,
|
||||||
raw_data=relationship_raw_data,
|
discovery_method="certificate_transparency_analysis"
|
||||||
discovery_method="certificate_transparency_analysis"
|
)
|
||||||
)
|
|
||||||
|
|
||||||
|
except json.JSONDecodeError as e:
|
||||||
|
self.logger.logger.error(f"Failed to parse JSON response from crt.sh: {e}")
|
||||||
|
except requests.exceptions.RequestException as e:
|
||||||
|
self.logger.logger.error(f"HTTP request to crt.sh failed: {e}")
|
||||||
|
|
||||||
|
|
||||||
return relationships
|
return relationships
|
||||||
|
|
||||||
def _find_shared_certificates(self, certs1: List[Dict[str, Any]], certs2: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
|
def _find_shared_certificates(self, certs1: List[Dict[str, Any]], certs2: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
|
||||||
@@ -541,15 +345,14 @@ class CrtShProvider(BaseProvider):
|
|||||||
'expires_soon_count': 0,
|
'expires_soon_count': 0,
|
||||||
'unique_issuers': [],
|
'unique_issuers': [],
|
||||||
'latest_certificate': None,
|
'latest_certificate': None,
|
||||||
'has_valid_cert': False,
|
'has_valid_cert': False
|
||||||
'certificate_details': [] # Always include empty list
|
|
||||||
}
|
}
|
||||||
|
|
||||||
valid_count = sum(1 for cert in certificates if cert.get('is_currently_valid'))
|
valid_count = sum(1 for cert in certificates if cert.get('is_currently_valid'))
|
||||||
expired_count = len(certificates) - valid_count
|
expired_count = len(certificates) - valid_count
|
||||||
expires_soon_count = sum(1 for cert in certificates if cert.get('expires_soon'))
|
expires_soon_count = sum(1 for cert in certificates if cert.get('expires_soon'))
|
||||||
|
|
||||||
# Get unique issuers (using parsed organization names)
|
# Get unique issuers
|
||||||
unique_issuers = list(set(cert.get('issuer_name') for cert in certificates if cert.get('issuer_name')))
|
unique_issuers = list(set(cert.get('issuer_name') for cert in certificates if cert.get('issuer_name')))
|
||||||
|
|
||||||
# Find the most recent certificate
|
# Find the most recent certificate
|
||||||
@@ -566,13 +369,6 @@ class CrtShProvider(BaseProvider):
|
|||||||
except Exception:
|
except Exception:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# Sort certificates by date for better display (newest first)
|
|
||||||
sorted_certificates = sorted(
|
|
||||||
certificates,
|
|
||||||
key=lambda c: self._get_certificate_sort_date(c),
|
|
||||||
reverse=True
|
|
||||||
)
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'total_certificates': len(certificates),
|
'total_certificates': len(certificates),
|
||||||
'valid_certificates': valid_count,
|
'valid_certificates': valid_count,
|
||||||
@@ -581,35 +377,9 @@ class CrtShProvider(BaseProvider):
|
|||||||
'unique_issuers': unique_issuers,
|
'unique_issuers': unique_issuers,
|
||||||
'latest_certificate': latest_cert,
|
'latest_certificate': latest_cert,
|
||||||
'has_valid_cert': valid_count > 0,
|
'has_valid_cert': valid_count > 0,
|
||||||
'certificate_details': sorted_certificates # Include full certificate details
|
'certificate_details': certificates # Full details for forensic analysis
|
||||||
}
|
}
|
||||||
|
|
||||||
def _get_certificate_sort_date(self, cert: Dict[str, Any]) -> datetime:
|
|
||||||
"""
|
|
||||||
Get a sortable date from certificate data for chronological ordering.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
cert: Certificate metadata dictionary
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Datetime object for sorting (falls back to epoch if parsing fails)
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
# Try not_before first (issue date)
|
|
||||||
if cert.get('not_before'):
|
|
||||||
return self._parse_certificate_date(cert['not_before'])
|
|
||||||
|
|
||||||
# Fall back to entry_timestamp if available
|
|
||||||
if cert.get('entry_timestamp'):
|
|
||||||
return self._parse_certificate_date(cert['entry_timestamp'])
|
|
||||||
|
|
||||||
# Last resort - return a very old date for certificates without dates
|
|
||||||
return datetime(1970, 1, 1, tzinfo=timezone.utc)
|
|
||||||
|
|
||||||
except Exception:
|
|
||||||
# If all parsing fails, return epoch
|
|
||||||
return datetime(1970, 1, 1, tzinfo=timezone.utc)
|
|
||||||
|
|
||||||
def _calculate_domain_relationship_confidence(self, domain1: str, domain2: str,
|
def _calculate_domain_relationship_confidence(self, domain1: str, domain2: str,
|
||||||
shared_certificates: List[Dict[str, Any]],
|
shared_certificates: List[Dict[str, Any]],
|
||||||
all_discovered_domains: Set[str]) -> float:
|
all_discovered_domains: Set[str]) -> float:
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
# dnsrecon/providers/dns_provider.py
|
# dnsrecon/providers/dns_provider.py
|
||||||
|
|
||||||
from dns import resolver, reversename
|
import dns.resolver
|
||||||
|
import dns.reversename
|
||||||
from typing import List, Dict, Any, Tuple
|
from typing import List, Dict, Any, Tuple
|
||||||
from .base_provider import BaseProvider
|
from .base_provider import BaseProvider
|
||||||
from utils.helpers import _is_valid_ip, _is_valid_domain
|
from utils.helpers import _is_valid_ip, _is_valid_domain
|
||||||
@@ -12,7 +13,7 @@ class DNSProvider(BaseProvider):
|
|||||||
Now uses session-specific configuration.
|
Now uses session-specific configuration.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, name=None, session_config=None):
|
def __init__(self, session_config=None):
|
||||||
"""Initialize DNS provider with session-specific configuration."""
|
"""Initialize DNS provider with session-specific configuration."""
|
||||||
super().__init__(
|
super().__init__(
|
||||||
name="dns",
|
name="dns",
|
||||||
@@ -22,7 +23,7 @@ class DNSProvider(BaseProvider):
|
|||||||
)
|
)
|
||||||
|
|
||||||
# Configure DNS resolver
|
# Configure DNS resolver
|
||||||
self.resolver = resolver.Resolver()
|
self.resolver = dns.resolver.Resolver()
|
||||||
self.resolver.timeout = 5
|
self.resolver.timeout = 5
|
||||||
self.resolver.lifetime = 10
|
self.resolver.lifetime = 10
|
||||||
#self.resolver.nameservers = ['127.0.0.1']
|
#self.resolver.nameservers = ['127.0.0.1']
|
||||||
@@ -50,7 +51,12 @@ class DNSProvider(BaseProvider):
|
|||||||
def query_domain(self, domain: str) -> List[Tuple[str, str, str, float, Dict[str, Any]]]:
|
def query_domain(self, domain: str) -> List[Tuple[str, str, str, float, Dict[str, Any]]]:
|
||||||
"""
|
"""
|
||||||
Query DNS records for the domain to discover relationships.
|
Query DNS records for the domain to discover relationships.
|
||||||
...
|
|
||||||
|
Args:
|
||||||
|
domain: Domain to investigate
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
List of relationships discovered from DNS analysis
|
||||||
"""
|
"""
|
||||||
if not _is_valid_domain(domain):
|
if not _is_valid_domain(domain):
|
||||||
return []
|
return []
|
||||||
@@ -59,15 +65,7 @@ class DNSProvider(BaseProvider):
|
|||||||
|
|
||||||
# Query all record types
|
# Query all record types
|
||||||
for record_type in ['A', 'AAAA', 'CNAME', 'MX', 'NS', 'SOA', 'TXT', 'SRV', 'CAA']:
|
for record_type in ['A', 'AAAA', 'CNAME', 'MX', 'NS', 'SOA', 'TXT', 'SRV', 'CAA']:
|
||||||
try:
|
relationships.extend(self._query_record(domain, record_type))
|
||||||
relationships.extend(self._query_record(domain, record_type))
|
|
||||||
except resolver.NoAnswer:
|
|
||||||
# This is not an error, just a confirmation that the record doesn't exist.
|
|
||||||
self.logger.logger.debug(f"No {record_type} record found for {domain}")
|
|
||||||
except Exception as e:
|
|
||||||
self.failed_requests += 1
|
|
||||||
self.logger.logger.debug(f"{record_type} record query failed for {domain}: {e}")
|
|
||||||
# Optionally, you might want to re-raise other, more serious exceptions.
|
|
||||||
|
|
||||||
return relationships
|
return relationships
|
||||||
|
|
||||||
@@ -89,7 +87,7 @@ class DNSProvider(BaseProvider):
|
|||||||
try:
|
try:
|
||||||
# Perform reverse DNS lookup
|
# Perform reverse DNS lookup
|
||||||
self.total_requests += 1
|
self.total_requests += 1
|
||||||
reverse_name = reversename.from_address(ip)
|
reverse_name = dns.reversename.from_address(ip)
|
||||||
response = self.resolver.resolve(reverse_name, 'PTR')
|
response = self.resolver.resolve(reverse_name, 'PTR')
|
||||||
self.successful_requests += 1
|
self.successful_requests += 1
|
||||||
|
|
||||||
@@ -121,14 +119,9 @@ class DNSProvider(BaseProvider):
|
|||||||
discovery_method="reverse_dns_lookup"
|
discovery_method="reverse_dns_lookup"
|
||||||
)
|
)
|
||||||
|
|
||||||
except resolver.NXDOMAIN:
|
|
||||||
self.failed_requests += 1
|
|
||||||
self.logger.logger.debug(f"Reverse DNS lookup failed for {ip}: NXDOMAIN")
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
self.failed_requests += 1
|
self.failed_requests += 1
|
||||||
self.logger.logger.debug(f"Reverse DNS lookup failed for {ip}: {e}")
|
self.logger.logger.debug(f"Reverse DNS lookup failed for {ip}: {e}")
|
||||||
# Re-raise the exception so the scanner can handle the failure
|
|
||||||
raise e
|
|
||||||
|
|
||||||
return relationships
|
return relationships
|
||||||
|
|
||||||
@@ -192,7 +185,5 @@ class DNSProvider(BaseProvider):
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
self.failed_requests += 1
|
self.failed_requests += 1
|
||||||
self.logger.logger.debug(f"{record_type} record query failed for {domain}: {e}")
|
self.logger.logger.debug(f"{record_type} record query failed for {domain}: {e}")
|
||||||
# Re-raise the exception so the scanner can handle it
|
|
||||||
raise e
|
|
||||||
|
|
||||||
return relationships
|
return relationships
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user