Update app.py
Browse files
app.py
CHANGED
@@ -113,6 +113,9 @@ def generate_large_synthetic_data(description, columns, num_rows=1000, rows_per_
|
|
113 |
|
114 |
for i in tqdm(range(num_iterations)):
|
115 |
generated_data = generate_synthetic_data(description, columns)
|
|
|
|
|
|
|
116 |
if "Error" in generated_data:
|
117 |
return generated_data
|
118 |
df_synthetic = process_generated_data(generated_data)
|
@@ -124,9 +127,26 @@ def generate_large_synthetic_data(description, columns, num_rows=1000, rows_per_
|
|
124 |
return pd.concat(data_frames, ignore_index=True)
|
125 |
|
126 |
def process_generated_data(csv_data):
|
127 |
-
|
128 |
-
|
129 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
130 |
|
131 |
# Streamlit app interface
|
132 |
st.title("Synthetic Data Generator")
|
|
|
113 |
|
114 |
for i in tqdm(range(num_iterations)):
|
115 |
generated_data = generate_synthetic_data(description, columns)
|
116 |
+
|
117 |
+
print("Generated Data:\n", generated_data) # Add this line to debug
|
118 |
+
|
119 |
if "Error" in generated_data:
|
120 |
return generated_data
|
121 |
df_synthetic = process_generated_data(generated_data)
|
|
|
127 |
return pd.concat(data_frames, ignore_index=True)
|
128 |
|
129 |
def process_generated_data(csv_data):
|
130 |
+
try:
|
131 |
+
# Check if the data is not empty and has valid content
|
132 |
+
if not csv_data.strip():
|
133 |
+
raise ValueError("Generated data is empty.")
|
134 |
+
|
135 |
+
data = StringIO(csv_data)
|
136 |
+
df = pd.read_csv(data)
|
137 |
+
print("DataFrame Shape:", df.shape)
|
138 |
+
print("DataFrame Head:\n", df.head())
|
139 |
+
|
140 |
+
|
141 |
+
# Check if the DataFrame is empty
|
142 |
+
if df.empty:
|
143 |
+
raise ValueError("Generated DataFrame is empty.")
|
144 |
+
|
145 |
+
return df
|
146 |
+
except Exception as e:
|
147 |
+
st.error(f"Error processing generated data: {e}")
|
148 |
+
return pd.DataFrame() # Return an empty DataFrame on error
|
149 |
+
|
150 |
|
151 |
# Streamlit app interface
|
152 |
st.title("Synthetic Data Generator")
|