|
@@ -1,8 +1,12 @@
|
|
|
import streamlit as st
|
|
import streamlit as st
|
|
|
import requests
|
|
import requests
|
|
|
|
|
+from ultralytics import YOLO
|
|
|
|
|
+import numpy as np
|
|
|
from PIL import Image
|
|
from PIL import Image
|
|
|
import io
|
|
import io
|
|
|
import base64
|
|
import base64
|
|
|
|
|
+import pandas as pd
|
|
|
|
|
+import plotly.express as px
|
|
|
|
|
|
|
|
# --- 1. Global Backend Check ---
|
|
# --- 1. Global Backend Check ---
|
|
|
API_BASE_URL = "http://localhost:8000"
|
|
API_BASE_URL = "http://localhost:8000"
|
|
@@ -16,6 +20,13 @@ def check_backend():
|
|
|
|
|
|
|
|
backend_active = check_backend()
|
|
backend_active = check_backend()
|
|
|
|
|
|
|
|
|
|
+# Load YOLO model locally for Analytical View
|
|
|
|
|
+@st.cache_resource
|
|
|
|
|
+def load_yolo():
|
|
|
|
|
+ return YOLO('best.pt')
|
|
|
|
|
+
|
|
|
|
|
+yolo_model = load_yolo()
|
|
|
|
|
+
|
|
|
if not backend_active:
|
|
if not backend_active:
|
|
|
st.error("⚠️ Backend API is offline!")
|
|
st.error("⚠️ Backend API is offline!")
|
|
|
st.info("Please start the backend server first (e.g., `python main.py`) to unlock AI features.")
|
|
st.info("Please start the backend server first (e.g., `python main.py`) to unlock AI features.")
|
|
@@ -69,7 +80,7 @@ with tab1:
|
|
|
# 1. Action Button (Centered and Prominent)
|
|
# 1. Action Button (Centered and Prominent)
|
|
|
st.write("##")
|
|
st.write("##")
|
|
|
_, col_btn, _ = st.columns([1, 2, 1])
|
|
_, col_btn, _ = st.columns([1, 2, 1])
|
|
|
- if col_btn.button("🔍 Run Ripeness Detection", type="primary", use_container_width=True):
|
|
|
|
|
|
|
+ if col_btn.button("🔍 Run Ripeness Detection", type="primary", width='stretch'):
|
|
|
with st.spinner("Processing Detections Locally..."):
|
|
with st.spinner("Processing Detections Locally..."):
|
|
|
files = {"file": (uploaded_file.name, uploaded_file.getvalue(), uploaded_file.type)}
|
|
files = {"file": (uploaded_file.name, uploaded_file.getvalue(), uploaded_file.type)}
|
|
|
res = requests.post(f"{API_BASE_URL}/analyze", files=files)
|
|
res = requests.post(f"{API_BASE_URL}/analyze", files=files)
|
|
@@ -79,44 +90,89 @@ with tab1:
|
|
|
else:
|
|
else:
|
|
|
st.error(f"Detection Failed: {res.text}")
|
|
st.error(f"Detection Failed: {res.text}")
|
|
|
|
|
|
|
|
- # 2. Results Layout
|
|
|
|
|
- if st.session_state.last_detection:
|
|
|
|
|
- st.divider()
|
|
|
|
|
- col1, col2 = st.columns([1.5, 1])
|
|
|
|
|
-
|
|
|
|
|
- with col1:
|
|
|
|
|
- st.image(uploaded_file, caption="Analyzed Image", use_container_width=True)
|
|
|
|
|
-
|
|
|
|
|
- with col2:
|
|
|
|
|
- data = st.session_state.last_detection
|
|
|
|
|
- with st.container(border=True):
|
|
|
|
|
- st.write("### 🏷️ Detection Results")
|
|
|
|
|
- if not data['detections']:
|
|
|
|
|
- st.warning("No Fresh Fruit Bunches detected.")
|
|
|
|
|
- else:
|
|
|
|
|
- for det in data['detections']:
|
|
|
|
|
- st.info(f"**{det['class']}** - {det['confidence']:.2%} confidence")
|
|
|
|
|
-
|
|
|
|
|
- # 3. Cloud Actions (Only if detections found)
|
|
|
|
|
- st.write("---")
|
|
|
|
|
- st.write("#### ✨ Cloud Archive")
|
|
|
|
|
- if st.button("🚀 Save to Atlas (Vectorize)", use_container_width=True):
|
|
|
|
|
- with st.spinner("Archiving..."):
|
|
|
|
|
- import json
|
|
|
|
|
- primary_det = data['detections'][0]
|
|
|
|
|
- payload = {"detection_data": json.dumps(primary_det)}
|
|
|
|
|
- files_cloud = {"file": (uploaded_file.name, uploaded_file.getvalue(), uploaded_file.type)}
|
|
|
|
|
-
|
|
|
|
|
- res_cloud = requests.post(f"{API_BASE_URL}/vectorize_and_store", files=files_cloud, data=payload)
|
|
|
|
|
-
|
|
|
|
|
- if res_cloud.status_code == 200:
|
|
|
|
|
- res_json = res_cloud.json()
|
|
|
|
|
- if res_json["status"] == "success":
|
|
|
|
|
- st.success(f"Archived! ID: `{res_json['record_id'][:8]}...`")
|
|
|
|
|
|
|
+ # 2. Results Layout
|
|
|
|
|
+ if st.session_state.last_detection:
|
|
|
|
|
+ st.divider()
|
|
|
|
|
+
|
|
|
|
|
+ # SIDE-BY-SIDE ANALYTICAL VIEW
|
|
|
|
|
+ col_left, col_right = st.columns(2)
|
|
|
|
|
+
|
|
|
|
|
+ with col_left:
|
|
|
|
|
+ st.image(uploaded_file, caption="Original Photo", width='stretch')
|
|
|
|
|
+
|
|
|
|
|
+ with col_right:
|
|
|
|
|
+ # Use the local model to plot the boxes directly
|
|
|
|
|
+ img = Image.open(uploaded_file)
|
|
|
|
|
+ results = yolo_model(img, conf=current_conf, agnostic_nms=True, iou=0.4)
|
|
|
|
|
+ annotated_img = results[0].plot() # Draws boxes/labels
|
|
|
|
|
+
|
|
|
|
|
+ # Convert BGR (OpenCV format) to RGB for Streamlit
|
|
|
|
|
+ annotated_img_rgb = annotated_img[:, :, ::-1]
|
|
|
|
|
+ st.image(annotated_img_rgb, caption="AI Analytical View (X-Ray)", width='stretch')
|
|
|
|
|
+
|
|
|
|
|
+ col1, col2 = st.columns([1.5, 1]) # Keep original col structure for summary below
|
|
|
|
|
+
|
|
|
|
|
+ with col2:
|
|
|
|
|
+ data = st.session_state.last_detection
|
|
|
|
|
+ with st.container(border=True):
|
|
|
|
|
+ st.write("### 🏷️ Detection Results")
|
|
|
|
|
+ if not data['detections']:
|
|
|
|
|
+ st.warning("No Fresh Fruit Bunches detected.")
|
|
|
|
|
+ else:
|
|
|
|
|
+ for det in data['detections']:
|
|
|
|
|
+ st.info(f"**{det['class']}** - {det['confidence']:.2%} confidence")
|
|
|
|
|
+
|
|
|
|
|
+ st.write("### 📊 Harvest Quality Mix")
|
|
|
|
|
+ # Convert industrial_summary dictionary to a DataFrame for charting
|
|
|
|
|
+ summary_df = pd.DataFrame(
|
|
|
|
|
+ list(data['industrial_summary'].items()),
|
|
|
|
|
+ columns=['Grade', 'Count']
|
|
|
|
|
+ )
|
|
|
|
|
+ # Filter out classes with 0 count for a cleaner chart
|
|
|
|
|
+ summary_df = summary_df[summary_df['Count'] > 0]
|
|
|
|
|
+
|
|
|
|
|
+ if not summary_df.empty:
|
|
|
|
|
+ # Create a Pie Chart to show the proportion of each grade
|
|
|
|
|
+ fig = px.pie(summary_df, values='Count', names='Grade',
|
|
|
|
|
+ color='Grade',
|
|
|
|
|
+ color_discrete_map={
|
|
|
|
|
+ 'Abnormal': '#ef4444', # Red
|
|
|
|
|
+ 'Empty_Bunch': '#94a3b8', # Gray
|
|
|
|
|
+ 'Ripe': '#22c55e', # Green
|
|
|
|
|
+ 'Underripe': '#eab308', # Yellow
|
|
|
|
|
+ 'Unripe': '#3b82f6', # Blue
|
|
|
|
|
+ 'Overripe': '#a855f7' # Purple
|
|
|
|
|
+ },
|
|
|
|
|
+ hole=0.4)
|
|
|
|
|
+ fig.update_layout(margin=dict(t=0, b=0, l=0, r=0), height=300)
|
|
|
|
|
+ st.plotly_chart(fig, width='stretch')
|
|
|
|
|
+
|
|
|
|
|
+ # High-Priority Health Alert
|
|
|
|
|
+ if data['industrial_summary'].get('Abnormal', 0) > 0:
|
|
|
|
|
+ st.error(f"🚨 CRITICAL: {data['industrial_summary']['Abnormal']} Abnormal Bunches Detected!")
|
|
|
|
|
+ if data['industrial_summary'].get('Empty_Bunch', 0) > 0:
|
|
|
|
|
+ st.warning(f"⚠️ ALERT: {data['industrial_summary']['Empty_Bunch']} Empty Bunches Detected.")
|
|
|
|
|
+
|
|
|
|
|
+ # 3. Cloud Actions (Only if detections found)
|
|
|
|
|
+ st.write("---")
|
|
|
|
|
+ st.write("#### ✨ Cloud Archive")
|
|
|
|
|
+ if st.button("🚀 Save to Atlas (Vectorize)", width='stretch'):
|
|
|
|
|
+ with st.spinner("Archiving..."):
|
|
|
|
|
+ import json
|
|
|
|
|
+ primary_det = data['detections'][0]
|
|
|
|
|
+ payload = {"detection_data": json.dumps(primary_det)}
|
|
|
|
|
+ files_cloud = {"file": (uploaded_file.name, uploaded_file.getvalue(), uploaded_file.type)}
|
|
|
|
|
+
|
|
|
|
|
+ res_cloud = requests.post(f"{API_BASE_URL}/vectorize_and_store", files=files_cloud, data=payload)
|
|
|
|
|
+
|
|
|
|
|
+ if res_cloud.status_code == 200:
|
|
|
|
|
+ res_json = res_cloud.json()
|
|
|
|
|
+ if res_json["status"] == "success":
|
|
|
|
|
+ st.success(f"Archived! ID: `{res_json['record_id'][:8]}...`")
|
|
|
|
|
+ else:
|
|
|
|
|
+ st.error(f"Cloud Error: {res_json['message']}")
|
|
|
else:
|
|
else:
|
|
|
- st.error(f"Cloud Error: {res_json['message']}")
|
|
|
|
|
- else:
|
|
|
|
|
- st.error("Failed to connect to cloud service")
|
|
|
|
|
|
|
+ st.error("Failed to connect to cloud service")
|
|
|
|
|
|
|
|
# --- Tab 2: Batch Processing ---
|
|
# --- Tab 2: Batch Processing ---
|
|
|
with tab2:
|
|
with tab2:
|
|
@@ -133,6 +189,31 @@ with tab2:
|
|
|
res_data = st.session_state.last_batch_results
|
|
res_data = st.session_state.last_batch_results
|
|
|
with st.container(border=True):
|
|
with st.container(border=True):
|
|
|
st.success(f"✅ Successfully processed {res_data['processed_count']} images.")
|
|
st.success(f"✅ Successfully processed {res_data['processed_count']} images.")
|
|
|
|
|
+
|
|
|
|
|
+ # Batch Summary Dashboard
|
|
|
|
|
+ st.write("### 📈 Batch Quality Overview")
|
|
|
|
|
+ batch_summary = res_data.get('industrial_summary', {})
|
|
|
|
|
+ if batch_summary:
|
|
|
|
|
+ sum_df = pd.DataFrame(list(batch_summary.items()), columns=['Grade', 'Count'])
|
|
|
|
|
+ sum_df = sum_df[sum_df['Count'] > 0]
|
|
|
|
|
+
|
|
|
|
|
+ b_col1, b_col2 = st.columns([1, 1])
|
|
|
|
|
+ with b_col1:
|
|
|
|
|
+ st.dataframe(sum_df, hide_index=True, width='stretch')
|
|
|
|
|
+ with b_col2:
|
|
|
|
|
+ if not sum_df.empty:
|
|
|
|
|
+ fig_batch = px.bar(sum_df, x='Grade', y='Count', color='Grade',
|
|
|
|
|
+ color_discrete_map={
|
|
|
|
|
+ 'Abnormal': '#ef4444',
|
|
|
|
|
+ 'Empty_Bunch': '#94a3b8',
|
|
|
|
|
+ 'Ripe': '#22c55e'
|
|
|
|
|
+ })
|
|
|
|
|
+ fig_batch.update_layout(margin=dict(t=0, b=0, l=0, r=0), height=200, showlegend=False)
|
|
|
|
|
+ st.plotly_chart(fig_batch, width='stretch')
|
|
|
|
|
+
|
|
|
|
|
+ if batch_summary.get('Abnormal', 0) > 0:
|
|
|
|
|
+ st.error(f"🚨 BATCH CRITICAL: {batch_summary['Abnormal']} Abnormal Bunches found in this batch!")
|
|
|
|
|
+
|
|
|
st.write("Generated Record IDs:")
|
|
st.write("Generated Record IDs:")
|
|
|
st.code(res_data['record_ids'])
|
|
st.code(res_data['record_ids'])
|
|
|
if st.button("Clear Results & Start New Batch"):
|
|
if st.button("Clear Results & Start New Batch"):
|