|
1 | 1 | use crate::models::{AppState, DataEntry, ExportRequest, Project}; |
2 | | -use crate::s3_helpers; |
3 | | -use bytes::Bytes; |
4 | | - |
5 | 2 | use axum::{ |
6 | 3 | extract::{Path, State}, |
7 | | - http::{StatusCode, header::LOCATION}, |
| 4 | + http::{header::LOCATION, StatusCode}, |
8 | 5 | response::{IntoResponse, Response}, |
9 | 6 | }; |
10 | 7 | use futures::stream::StreamExt; |
11 | | -use sqlx::Row; |
| 8 | +use tokio::io::AsyncWriteExt; |
12 | 9 |
|
13 | 10 | pub async fn export( |
14 | 11 | State(state): State<AppState>, |
@@ -42,82 +39,144 @@ pub async fn export( |
42 | 39 |
|
43 | 40 | let s3_key = format!("exports/{}.json", token); |
44 | 41 |
|
45 | | - let file_exists = s3_helpers::check_file_exists(&state.s3_bucket, &s3_key) |
| 42 | + let file_exists = state.s3_bucket.object_exists(&s3_key).await.map_err(|e| { |
| 43 | + println!("Error checking S3 file existence: {:?}", e); |
| 44 | + StatusCode::INTERNAL_SERVER_ERROR |
| 45 | + })?; |
| 46 | + |
| 47 | + if !file_exists { |
| 48 | + stream_export_to_s3(&state, export_request.project_id, &s3_key).await?; |
| 49 | + } |
| 50 | + |
| 51 | + let presigned_url = state |
| 52 | + .s3_bucket |
| 53 | + .presign_get(&s3_key, 300, None) |
46 | 54 | .await |
47 | 55 | .map_err(|e| { |
48 | | - println!("Error checking S3 file existence: {:?}", e); |
| 56 | + println!("Error generating presigned URL: {:?}", e); |
49 | 57 | StatusCode::INTERNAL_SERVER_ERROR |
50 | 58 | })?; |
51 | 59 |
|
52 | | - if !file_exists { |
53 | | - let export_data = generate_export_data(&state, export_request.project_id).await?; |
| 60 | + Ok((StatusCode::FOUND, [(LOCATION, presigned_url.as_str())]).into_response()) |
| 61 | +} |
54 | 62 |
|
55 | | - s3_helpers::upload_file(&state.s3_bucket, &s3_key, export_data) |
56 | | - .await |
57 | | - .map_err(|e| { |
58 | | - println!("Error uploading to S3: {:?}", e); |
59 | | - StatusCode::INTERNAL_SERVER_ERROR |
60 | | - })?; |
| 63 | +fn indent_json(json: &[u8], spaces: usize) -> Vec<u8> { |
| 64 | + let mut result = Vec::with_capacity(json.len() + (json.len() >> 3)); |
| 65 | + |
| 66 | + for &byte in json { |
| 67 | + result.push(byte); |
| 68 | + if byte == b'\n' { |
| 69 | + result.extend(std::iter::repeat_n(b' ', spaces)); |
| 70 | + } |
| 71 | + } |
| 72 | + |
| 73 | + while result.last() == Some(&b' ') { |
| 74 | + result.pop(); |
61 | 75 | } |
62 | 76 |
|
63 | | - let presigned_url = s3_helpers::generate_presigned_url(&state.s3_bucket, &s3_key, 300) |
| 77 | + result |
| 78 | +} |
| 79 | + |
| 80 | +async fn stream_export_to_s3( |
| 81 | + state: &AppState, |
| 82 | + project_id: sqlx::types::Uuid, |
| 83 | + s3_key: &str, |
| 84 | +) -> Result<(), StatusCode> { |
| 85 | + let (writer, mut reader) = tokio::io::duplex(128 * 1024); |
| 86 | + |
| 87 | + let pool = state.pool.clone(); |
| 88 | + let s3_bucket = state.s3_bucket.clone(); |
| 89 | + let s3_key_owned = s3_key.to_string(); |
| 90 | + |
| 91 | + let writer_handle = tokio::spawn(async move { |
| 92 | + let project = sqlx::query_as::<_, Project>( |
| 93 | + "SELECT id, name, token, slug, private, template_id, created_at, owner_id FROM project WHERE id = $1" |
| 94 | + ) |
| 95 | + .bind(project_id) |
| 96 | + .fetch_optional(&pool) |
| 97 | + .await |
| 98 | + .map_err(std::io::Error::other)? |
| 99 | + .ok_or_else(|| std::io::Error::new(std::io::ErrorKind::NotFound, "Project not found"))?; |
| 100 | + |
| 101 | + write_export_json(writer, project, &pool, project_id).await |
| 102 | + }); |
| 103 | + |
| 104 | + let upload_handle = tokio::spawn(async move { |
| 105 | + s3_bucket |
| 106 | + .put_object_stream(&mut reader, &s3_key_owned) |
| 107 | + .await |
| 108 | + .map_err(std::io::Error::other) |
| 109 | + }); |
| 110 | + |
| 111 | + writer_handle |
64 | 112 | .await |
65 | 113 | .map_err(|e| { |
66 | | - println!("Error generating presigned URL: {:?}", e); |
| 114 | + println!("Writer task failed: {:?}", e); |
| 115 | + StatusCode::INTERNAL_SERVER_ERROR |
| 116 | + })? |
| 117 | + .map_err(|e| { |
| 118 | + println!("Error writing JSON: {:?}", e); |
67 | 119 | StatusCode::INTERNAL_SERVER_ERROR |
68 | 120 | })?; |
69 | 121 |
|
70 | | - Ok((StatusCode::FOUND, [(LOCATION, presigned_url.as_str())]).into_response()) |
| 122 | + upload_handle |
| 123 | + .await |
| 124 | + .map_err(|e| { |
| 125 | + println!("Upload task failed: {:?}", e); |
| 126 | + StatusCode::INTERNAL_SERVER_ERROR |
| 127 | + })? |
| 128 | + .map_err(|e| { |
| 129 | + println!("Error uploading to S3: {:?}", e); |
| 130 | + StatusCode::INTERNAL_SERVER_ERROR |
| 131 | + })?; |
| 132 | + |
| 133 | + Ok(()) |
71 | 134 | } |
72 | 135 |
|
73 | | -async fn generate_export_data( |
74 | | - state: &AppState, |
| 136 | +async fn write_export_json<W>( |
| 137 | + mut writer: W, |
| 138 | + project: Project, |
| 139 | + pool: &sqlx::PgPool, |
75 | 140 | project_id: sqlx::types::Uuid, |
76 | | -) -> Result<Bytes, StatusCode> { |
77 | | - let project = sqlx::query_as::<_, Project>( |
78 | | - "SELECT id, name, token, slug, private, template_id, created_at, owner_id FROM project WHERE id = $1" |
79 | | - ) |
80 | | - .bind(project_id) |
81 | | - .fetch_optional(&state.pool) |
82 | | - .await |
83 | | - .map_err(|e| { |
84 | | - println!("Error while fetching project: {:?}", e); |
85 | | - StatusCode::INTERNAL_SERVER_ERROR |
86 | | - })? |
87 | | - .ok_or(StatusCode::NOT_FOUND)?; |
| 141 | +) -> Result<(), std::io::Error> |
| 142 | +where |
| 143 | + W: tokio::io::AsyncWrite + Unpin, |
| 144 | +{ |
| 145 | + let mut buffer = Vec::with_capacity(131072); |
| 146 | + |
| 147 | + writer.write_all(b"{\n \"project\": ").await?; |
| 148 | + |
| 149 | + serde_json::to_writer_pretty(&mut buffer, &project).map_err(std::io::Error::other)?; |
88 | 150 |
|
89 | | - let mut data_entries = Vec::new(); |
90 | | - let mut row_stream = sqlx::query( |
| 151 | + let indented = indent_json(&buffer, 2); |
| 152 | + writer.write_all(&indented).await?; |
| 153 | + |
| 154 | + writer.write_all(b",\n \"data_entries\": [\n").await?; |
| 155 | + |
| 156 | + let mut entries_stream = sqlx::query_as::<_, DataEntry>( |
91 | 157 | "SELECT data, created_at FROM data_entries WHERE project_id = $1 ORDER BY created_at DESC", |
92 | 158 | ) |
93 | 159 | .bind(project_id) |
94 | | - .fetch(&state.pool); |
95 | | - |
96 | | - while let Some(row) = row_stream.next().await { |
97 | | - match row { |
98 | | - Ok(row) => { |
99 | | - let data_entry = DataEntry { |
100 | | - data: row.try_get("data").ok(), |
101 | | - created_at: row.get("created_at"), |
102 | | - }; |
103 | | - data_entries.push(data_entry); |
104 | | - } |
105 | | - Err(e) => { |
106 | | - println!("Error while streaming data entry: {:?}", e); |
107 | | - return Err(StatusCode::INTERNAL_SERVER_ERROR); |
108 | | - } |
| 160 | + .fetch(pool); |
| 161 | + |
| 162 | + let mut first = true; |
| 163 | + while let Some(entry_result) = entries_stream.next().await { |
| 164 | + let entry = entry_result.map_err(std::io::Error::other)?; |
| 165 | + |
| 166 | + if !first { |
| 167 | + writer.write_all(b",\n").await?; |
109 | 168 | } |
110 | | - } |
| 169 | + first = false; |
111 | 170 |
|
112 | | - let export_data = serde_json::json!({ |
113 | | - "project": project, |
114 | | - "data_entries": data_entries |
115 | | - }); |
| 171 | + writer.write_all(b" ").await?; |
116 | 172 |
|
117 | | - let json_string = serde_json::to_string_pretty(&export_data).map_err(|e| { |
118 | | - println!("Error while serializing export data: {:?}", e); |
119 | | - StatusCode::INTERNAL_SERVER_ERROR |
120 | | - })?; |
| 173 | + buffer.clear(); |
| 174 | + serde_json::to_writer_pretty(&mut buffer, &entry).map_err(std::io::Error::other)?; |
| 175 | + |
| 176 | + let indented = indent_json(&buffer, 4); |
| 177 | + writer.write_all(&indented).await?; |
| 178 | + } |
121 | 179 |
|
122 | | - Ok(Bytes::from(json_string)) |
| 180 | + writer.write_all(b"\n ]\n}\n").await?; |
| 181 | + writer.shutdown().await |
123 | 182 | } |
0 commit comments