π‘ ExamplesΒΆ
Real-world examples demonstrating Ryomaβs capabilities across different use cases and industries.
π οΈ Setup Helper FunctionΒΆ
All examples below use this helper function to set up the agent:
from ryoma_ai.services import AgentBuilder, DataSourceService
from ryoma_ai.infrastructure.datasource_repository import StoreBasedDataSourceRepository
from langchain_core.stores import InMemoryStore
from ryoma_data.sql import DataSource
def create_sql_agent(datasource: DataSource, model: str = "gpt-4", mode: str = "enhanced"):
"""Helper function to create a SQL agent with a datasource."""
store = InMemoryStore()
repo = StoreBasedDataSourceRepository(store)
datasource_service = DataSourceService(repo)
datasource_service.add_datasource(datasource)
builder = AgentBuilder(datasource_service)
return builder.build_sql_agent(model=model, mode=mode)
πͺ E-commerce AnalyticsΒΆ
Customer Segmentation AnalysisΒΆ
from ryoma_data.sql import DataSource
# Connect to e-commerce database
datasource = DataSource(
"postgres",
host="localhost",
database="ecommerce",
user="user",
password="password"
)
agent = create_sql_agent(datasource, model="gpt-4", mode="enhanced")
# Complex customer analysis
response = agent.stream("""
Segment customers based on their purchase behavior:
1. High-value customers (>$1000 total purchases)
2. Frequent buyers (>10 orders)
3. Recent customers (first purchase in last 30 days)
4. At-risk customers (no purchase in 90+ days)
Show the count and average order value for each segment.
""")
print(response)
Product Performance DashboardΒΆ
# Multi-dimensional product analysis
response = agent.stream("""
Create a product performance report showing:
- Top 10 products by revenue this quarter
- Products with declining sales (comparing this month vs last month)
- Inventory turnover rate by category
- Seasonal trends for each product category
Include percentage changes and rank products by profitability.
""")
π Financial AnalysisΒΆ
Revenue ForecastingΒΆ
from ryoma_data.sql import DataSource
# Connect to financial data warehouse
datasource = DataSource(
"snowflake",
account="your-account",
database="FINANCE_DW",
user="user",
password="password"
)
agent = create_sql_agent(datasource, model="gpt-4", mode="reforce")
# Complex financial analysis
response = agent.stream("""
Analyze monthly recurring revenue (MRR) trends:
1. Calculate MRR growth rate for the last 12 months
2. Identify churn impact on revenue
3. Segment MRR by customer size (SMB, Mid-market, Enterprise)
4. Forecast next quarter MRR based on current trends
5. Show the contribution of new vs expansion revenue
Include confidence intervals for the forecast.
""")
Risk AssessmentΒΆ
# Credit risk analysis
response = agent.stream("""
Generate a credit risk report:
- Customers with overdue payments >30 days
- Payment pattern analysis by customer segment
- Default probability scoring based on historical data
- Recommended credit limits for new customers
- Geographic risk distribution
Rank customers by risk score and suggest actions.
""")
π₯ Healthcare AnalyticsΒΆ
Patient Outcome AnalysisΒΆ
from ryoma_data.sql import DataSource
# Connect to healthcare data
datasource = DataSource(
"bigquery",
project_id="healthcare-analytics",
dataset_id="patient_data"
)
# Create agent with enhanced mode for healthcare analysis
agent = create_sql_agent(datasource, model="gpt-4", mode="enhanced")
# HIPAA-compliant analysis
response = agent.stream("""
Analyze patient readmission patterns (anonymized data):
1. 30-day readmission rates by department
2. Common diagnoses leading to readmissions
3. Seasonal patterns in emergency visits
4. Average length of stay by condition type
5. Resource utilization efficiency metrics
Exclude any personally identifiable information.
""")
π Manufacturing OperationsΒΆ
Supply Chain OptimizationΒΆ
from ryoma_data.sql import DataSource
# Manufacturing database with IoT sensor data
datasource = DataSource(
"postgres",
host="localhost",
database="manufacturing",
user="user",
password="password"
)
agent = create_sql_agent(datasource, model="gpt-4", mode="enhanced")
# Complex supply chain analysis
response = agent.stream("""
Optimize our supply chain operations:
1. Identify bottlenecks in production lines
2. Calculate optimal inventory levels by component
3. Predict maintenance needs based on sensor data
4. Analyze supplier performance and delivery times
5. Recommend production schedule adjustments
Focus on reducing costs while maintaining quality standards.
""")
π± SaaS Product AnalyticsΒΆ
User Engagement AnalysisΒΆ
from ryoma_ai.agent.pandas_agent import PandasAgent
import pandas as pd
# Load user activity data
df = pd.read_csv('user_activity.csv')
agent = PandasAgent("gpt-4")
agent.add_dataframe(df, df_id="user_activity")
# Comprehensive user analysis
response = agent.stream("""
Analyze user engagement patterns:
1. Daily/weekly/monthly active users trends
2. Feature adoption rates and user journey analysis
3. Cohort retention analysis by signup month
4. Identify power users vs casual users
5. Churn prediction based on activity patterns
Create actionable insights for product team.
""")
A/B Test AnalysisΒΆ
# A/B test results analysis
test_data = pd.read_csv('ab_test_results.csv')
agent.add_dataframe(test_data, df_id="ab_test")
response = agent.stream("""
Analyze A/B test results for new checkout flow:
1. Statistical significance of conversion rate difference
2. Segment analysis (mobile vs desktop, new vs returning users)
3. Revenue impact calculation
4. Confidence intervals and p-values
5. Recommendation on whether to ship the new feature
Include both statistical and business impact analysis.
""")
π Educational Data AnalysisΒΆ
Student Performance InsightsΒΆ
from ryoma_data.sql import DataSource
# Educational database
datasource = DataSource(
"postgres",
host="localhost",
database="education",
user="user",
password="password"
)
agent = create_sql_agent(datasource, model="gpt-4", mode="enhanced")
response = agent.stream("""
Analyze student performance and learning outcomes:
1. Grade distribution by subject and semester
2. Correlation between attendance and performance
3. Identify at-risk students early warning indicators
4. Course completion rates and dropout patterns
5. Teacher effectiveness metrics
Provide recommendations for improving student success.
""")
π Multi-Database AnalysisΒΆ
Cross-Platform IntegrationΒΆ
from ryoma_data.sql import DataSource
# For multi-database analysis, create separate agents for each datasource
# or use a data warehouse that consolidates data
# Option 1: Analyze each database separately
sales_ds = DataSource("postgres", host="localhost", database="sales", user="user", password="password")
sales_agent = create_sql_agent(sales_ds, model="gpt-4", mode="enhanced")
marketing_ds = DataSource("snowflake", account="account", database="MARKETING", user="user", password="password")
marketing_agent = create_sql_agent(marketing_ds, model="gpt-4", mode="enhanced")
# Query each database
sales_response = sales_agent.stream("Get sales conversion rates by channel")
marketing_response = marketing_agent.stream("Get marketing campaign performance metrics")
# Option 2: Use a consolidated data warehouse
dw_ds = DataSource("snowflake", account="account", database="DATA_WAREHOUSE", user="user", password="password")
dw_agent = create_sql_agent(dw_ds, model="gpt-4", mode="enhanced")
# Cross-database analysis from consolidated warehouse
response = dw_agent.stream("""
Analyze the complete customer journey:
1. Marketing campaign effectiveness
2. Sales conversion rates
3. Customer lifetime value calculation
4. Attribution modeling across channels
5. ROI by marketing channel and campaign
""")
π Advanced Profiling ExamplesΒΆ
Database Metadata ExplorationΒΆ
from ryoma_data.sql import DataSource
# Connect to database
datasource = DataSource(
"postgres",
host="localhost",
database="production",
user="user",
password="password"
)
# Get catalog metadata
catalog = datasource.get_catalog()
print("π Database Catalog:")
for schema in catalog.schemas:
print(f"\nSchema: {schema.schema_name}")
for table in schema.tables:
print(f" Table: {table.table_name}")
for col in table.columns:
print(f" - {col.name}: {col.type}")
Table Search and InspectionΒΆ
# Search for tables matching a pattern
tables = datasource.search_tables(pattern="*customer*")
print(f"π Found {len(tables)} customer-related tables")
# Search for columns across all tables
email_columns = datasource.search_columns(pattern="*email*")
print(f"π§ Found {len(email_columns)} email-related columns")
# Get detailed table inspection with sample data
table_info = datasource.inspect_table("customers", include_sample_data=True, sample_limit=5)
print(f"π Columns in customers table: {[col.name for col in table_info.columns]}")
π― Best Practices from ExamplesΒΆ
1. Use Appropriate Agent ModesΒΆ
Enhanced mode for production analytics
ReFoRCE mode for complex, multi-step analysis
Basic mode for simple queries and development
2. Configure Safety for Your Use CaseΒΆ
Healthcare: Strict PII protection
Finance: Audit logging and access controls
E-commerce: Row limits for large datasets
3. Optimize Profiling SettingsΒΆ
Large datasets: Higher sample sizes
Real-time systems: Cached profiles
Development: Smaller samples for speed
4. Structure Complex QueriesΒΆ
Break down multi-part questions
Specify desired output format
Include business context and constraints
5. Monitor and IterateΒΆ
Track query performance
Refine prompts based on results
Use profiling data to improve accuracy
π Next StepsΒΆ
Ready to implement these patterns in your own projects?
Advanced Setup - Production configuration
API Reference - Complete method documentation
Architecture Guide - Deep dive into internals