rewrite generator
This commit is contained in:
32
src/generator/column.rs
Normal file
32
src/generator/column.rs
Normal file
@@ -0,0 +1,32 @@
|
||||
use color_eyre::{eyre::ContextCompat, Result};
|
||||
use sea_schema::sea_query::{ColumnDef, ColumnSpec, ColumnType, IndexCreateStatement};
|
||||
#[derive(Clone, Debug)]
|
||||
pub struct Column {
|
||||
pub name: String,
|
||||
pub col_type: ColumnType,
|
||||
pub attrs: Vec<ColumnSpec>,
|
||||
}
|
||||
|
||||
impl Column {
|
||||
pub fn new(column: ColumnDef, index: Option<IndexCreateStatement>) -> Result<Self> {
|
||||
let name = column.get_column_name();
|
||||
let col_type = column
|
||||
.get_column_type()
|
||||
.context("Unable to get column type")?
|
||||
.clone();
|
||||
let mut attrs = column.get_column_spec().clone();
|
||||
if let Some(index) = index {
|
||||
if index.is_unique_key() {
|
||||
attrs.push(ColumnSpec::UniqueKey)
|
||||
}
|
||||
if index.is_primary_key() {
|
||||
attrs.push(ColumnSpec::PrimaryKey);
|
||||
}
|
||||
}
|
||||
Ok(Column {
|
||||
name: name.to_string(),
|
||||
col_type,
|
||||
attrs: attrs.to_vec(),
|
||||
})
|
||||
}
|
||||
}
|
||||
150
src/generator/discover.rs
Normal file
150
src/generator/discover.rs
Normal file
@@ -0,0 +1,150 @@
|
||||
use core::time;
|
||||
|
||||
use color_eyre::eyre::{eyre, Context, ContextCompat, Report, Result};
|
||||
use sea_schema::sea_query::TableCreateStatement;
|
||||
use url::Url;
|
||||
|
||||
use crate::config::Config;
|
||||
pub async fn get_tables(
|
||||
database_url: String,
|
||||
config: &Config,
|
||||
) -> Result<(Option<String>, Vec<TableCreateStatement>)> {
|
||||
let url = Url::parse(&database_url)?;
|
||||
|
||||
tracing::trace!(?url);
|
||||
|
||||
let is_sqlite = url.scheme() == "sqlite";
|
||||
let filter_tables = config.sea_orm.entity.tables.get_filter();
|
||||
|
||||
let database_name: &str = (if !is_sqlite {
|
||||
let database_name = url
|
||||
.path_segments()
|
||||
.context("No database name as part of path")?
|
||||
.next()
|
||||
.context("No database name as part of path")?;
|
||||
|
||||
if database_name.is_empty() {
|
||||
return Err(eyre!("Database path name is empty"));
|
||||
}
|
||||
Ok::<&str, Report>(database_name)
|
||||
} else {
|
||||
Ok(Default::default())
|
||||
})?;
|
||||
|
||||
let (schema_name, table_stmts) = match url.scheme() {
|
||||
"mysql" => {
|
||||
use sea_schema::mysql::discovery::SchemaDiscovery;
|
||||
use sqlx::MySql;
|
||||
|
||||
tracing::info!("Connecting to MySQL");
|
||||
let connection = sqlx_connect::<MySql>(
|
||||
config.db.max_connections,
|
||||
config.db.acquire_timeout,
|
||||
url.as_str(),
|
||||
None,
|
||||
)
|
||||
.await?;
|
||||
|
||||
tracing::info!("Discovering schema");
|
||||
let schema_discovery = SchemaDiscovery::new(connection, database_name);
|
||||
let schema = schema_discovery.discover().await?;
|
||||
let table_stmts = schema
|
||||
.tables
|
||||
.into_iter()
|
||||
.filter(|schema| filter_tables(&schema.info.name))
|
||||
// .filter(|schema| filter_hidden_tables(&schema.info.name))
|
||||
// .filter(|schema| filter_skip_tables(&schema.info.name))
|
||||
.map(|schema| schema.write())
|
||||
.collect();
|
||||
(None, table_stmts)
|
||||
}
|
||||
"sqlite" => {
|
||||
use sea_schema::sqlite::discovery::SchemaDiscovery;
|
||||
use sqlx::Sqlite;
|
||||
|
||||
tracing::info!("Connecting to SQLite");
|
||||
let connection = sqlx_connect::<Sqlite>(
|
||||
config.db.max_connections,
|
||||
config.db.acquire_timeout,
|
||||
url.as_str(),
|
||||
None,
|
||||
)
|
||||
.await?;
|
||||
|
||||
tracing::info!("Discovering schema");
|
||||
let schema_discovery = SchemaDiscovery::new(connection);
|
||||
let schema = schema_discovery
|
||||
.discover()
|
||||
.await?
|
||||
.merge_indexes_into_table();
|
||||
let table_stmts = schema
|
||||
.tables
|
||||
.into_iter()
|
||||
.filter(|schema| filter_tables(&schema.name))
|
||||
// .filter(|schema| filter_hidden_tables(&schema.name))
|
||||
// .filter(|schema| filter_skip_tables(&schema.name))
|
||||
.map(|schema| schema.write())
|
||||
.collect();
|
||||
(None, table_stmts)
|
||||
}
|
||||
"postgres" | "potgresql" => {
|
||||
use sea_schema::postgres::discovery::SchemaDiscovery;
|
||||
use sqlx::Postgres;
|
||||
|
||||
tracing::info!("Connecting to Postgres");
|
||||
let schema = &config.db.database_schema.as_deref().unwrap_or("public");
|
||||
let connection = sqlx_connect::<Postgres>(
|
||||
config.db.max_connections,
|
||||
config.db.acquire_timeout,
|
||||
url.as_str(),
|
||||
Some(schema),
|
||||
)
|
||||
.await?;
|
||||
tracing::info!("Discovering schema");
|
||||
let schema_discovery = SchemaDiscovery::new(connection, schema);
|
||||
let schema = schema_discovery.discover().await?;
|
||||
tracing::info!(?schema);
|
||||
let table_stmts = schema
|
||||
.tables
|
||||
.into_iter()
|
||||
.filter(|schema| filter_tables(&schema.info.name))
|
||||
// .filter(|schema| filter_hidden_tables(&schema.info.name))
|
||||
// .filter(|schema| filter_skip_tables(&schema.info.name))
|
||||
.map(|schema| schema.write())
|
||||
.collect();
|
||||
(config.db.database_schema.clone(), table_stmts)
|
||||
}
|
||||
_ => unimplemented!("{} is not supported", url.scheme()),
|
||||
};
|
||||
tracing::info!("Schema discovered");
|
||||
|
||||
Ok((schema_name, table_stmts))
|
||||
}
|
||||
async fn sqlx_connect<DB>(
|
||||
max_connections: u32,
|
||||
acquire_timeout: u64,
|
||||
url: &str,
|
||||
schema: Option<&str>,
|
||||
) -> Result<sqlx::Pool<DB>>
|
||||
where
|
||||
DB: sqlx::Database,
|
||||
for<'a> &'a mut <DB as sqlx::Database>::Connection: sqlx::Executor<'a>,
|
||||
{
|
||||
let mut pool_options = sqlx::pool::PoolOptions::<DB>::new()
|
||||
.max_connections(max_connections)
|
||||
.acquire_timeout(time::Duration::from_secs(acquire_timeout));
|
||||
// Set search_path for Postgres, E.g. Some("public") by default
|
||||
// MySQL & SQLite connection initialize with schema `None`
|
||||
if let Some(schema) = schema {
|
||||
let sql = format!("SET search_path = '{schema}'");
|
||||
pool_options = pool_options.after_connect(move |conn, _| {
|
||||
let sql = sql.clone();
|
||||
Box::pin(async move {
|
||||
sqlx::Executor::execute(conn, sql.as_str())
|
||||
.await
|
||||
.map(|_| ())
|
||||
})
|
||||
});
|
||||
}
|
||||
pool_options.connect(url).await.map_err(Into::into)
|
||||
}
|
||||
95
src/generator/file.rs
Normal file
95
src/generator/file.rs
Normal file
@@ -0,0 +1,95 @@
|
||||
use std::path::PathBuf;
|
||||
|
||||
use crate::config::Config;
|
||||
|
||||
use super::table::Table;
|
||||
use color_eyre::Result;
|
||||
use handlebars::Handlebars;
|
||||
use prettytable::{format, row, Table as PTable};
|
||||
use sea_orm_codegen::OutputFile;
|
||||
use serde::Serialize;
|
||||
use tokio::fs;
|
||||
const HEADER: &str = r#"== Schema Information"#;
|
||||
const COMMENTHEAD: &str = r#"/*"#;
|
||||
const COMMENTBODY: &str = r#" *"#;
|
||||
const COMMENTTAIL: &str = r#"*/"#;
|
||||
|
||||
#[derive(Debug, Clone)]
|
||||
pub struct FileGenerator {
|
||||
filename: String,
|
||||
table: Table,
|
||||
}
|
||||
#[derive(Debug, Clone, Serialize)]
|
||||
pub struct FileContext {
|
||||
has_prelude: bool,
|
||||
prelude: String,
|
||||
}
|
||||
|
||||
impl FileGenerator {
|
||||
pub fn new(table: Table) -> Result<FileGenerator> {
|
||||
let filename = format!("{}.rs", table.name);
|
||||
Ok(FileGenerator { table, filename })
|
||||
}
|
||||
pub async fn build_file<'a>(
|
||||
&self,
|
||||
config: &Config,
|
||||
handlebars: &'a Handlebars<'a>,
|
||||
) -> Result<OutputFile> {
|
||||
let filepath = config.output.path.join(&self.filename);
|
||||
Ok(OutputFile {
|
||||
name: filepath.to_str().unwrap().to_string(),
|
||||
content: self.generate_file(config, handlebars).await?,
|
||||
})
|
||||
}
|
||||
pub async fn generate_file<'a>(
|
||||
&self,
|
||||
config: &Config,
|
||||
handlebars: &'a Handlebars<'a>,
|
||||
) -> Result<String> {
|
||||
let filepath = config.output.path.join(&self.filename);
|
||||
let file_context = FileContext {
|
||||
has_prelude: false,
|
||||
prelude: String::new(),
|
||||
};
|
||||
let generated_header = self.generate_header(config).await?;
|
||||
if filepath.exists() {
|
||||
let mut file_content = fs::read_to_string(filepath).await?;
|
||||
Ok(file_content)
|
||||
} else {
|
||||
let content = handlebars.render("model", &file_context)?;
|
||||
Ok(format!("{}{}", generated_header, content))
|
||||
}
|
||||
}
|
||||
pub async fn generate_header(&self, config: &Config) -> Result<String> {
|
||||
let mut column_info_table = PTable::new();
|
||||
let format = format::FormatBuilder::new()
|
||||
.column_separator(' ')
|
||||
.borders(' ')
|
||||
.separators(
|
||||
&[
|
||||
format::LinePosition::Bottom,
|
||||
format::LinePosition::Title,
|
||||
// format::LinePosition::Top,
|
||||
],
|
||||
format::LineSeparator::default(),
|
||||
)
|
||||
.padding(1, 1)
|
||||
.build();
|
||||
column_info_table.set_format(format);
|
||||
column_info_table.set_titles(row!["Name", "Type", "RustType", "Attributes"]);
|
||||
// let indexes = table.get_indexes();
|
||||
// tracing::info!(?indexes);
|
||||
// for column in table.get_columns() {
|
||||
// let name = column.get_column_name();
|
||||
// if let Some(column_type) = column.get_column_type() {
|
||||
// let column_type_rust =
|
||||
// type_to_rust_string(column_type, config.sea_orm.entity.date_time_crate.clone());
|
||||
// let column_type =
|
||||
// type_to_string(column_type, config.sea_orm.entity.date_time_crate.clone());
|
||||
// let attrs = attrs_to_string(column.get_column_spec());
|
||||
// ptable.add_row(row![name, column_type, column_type_rust, attrs]);
|
||||
// }
|
||||
// }
|
||||
Ok(String::new())
|
||||
}
|
||||
}
|
||||
213
src/generator/mod.rs
Normal file
213
src/generator/mod.rs
Normal file
@@ -0,0 +1,213 @@
|
||||
pub mod column;
|
||||
pub mod discover;
|
||||
pub mod file;
|
||||
pub mod table;
|
||||
|
||||
use core::time;
|
||||
|
||||
use crate::{config::DateTimeCrate, Config};
|
||||
use color_eyre::{
|
||||
eyre::{eyre, ContextCompat, Report},
|
||||
Result,
|
||||
};
|
||||
use comment_parser::{CommentParser, Event};
|
||||
use handlebars::Handlebars;
|
||||
use prettytable::{format, row};
|
||||
use sea_orm_codegen::OutputFile;
|
||||
use sea_schema::sea_query::{self, ColumnSpec, ColumnType, StringLen, TableCreateStatement};
|
||||
use table::Table;
|
||||
use tokio::{fs, task::JoinSet};
|
||||
use url::Url;
|
||||
|
||||
pub async fn update_files<'a>(
|
||||
tables: Vec<TableCreateStatement>,
|
||||
config: Config,
|
||||
handlebars: &'a Handlebars<'a>,
|
||||
) -> Result<Vec<OutputFile>> {
|
||||
let tables = tables
|
||||
.into_iter()
|
||||
.map(|table| Table::new(table.clone(), config.clone()))
|
||||
.collect::<Result<Vec<Table>>>()?;
|
||||
let mut files = Vec::new();
|
||||
for table in tables {
|
||||
let comment = file::FileGenerator::new(table)?;
|
||||
let file = comment.build_file(&config, &handlebars).await?;
|
||||
files.push(file);
|
||||
}
|
||||
Ok(Vec::new())
|
||||
}
|
||||
|
||||
// pub async fn generate_models(
|
||||
// tables: Vec<TableCreateStatement>,
|
||||
// config: Config,
|
||||
// ) -> Result<Vec<OutputFile>> {
|
||||
// tracing::debug!(?tables);
|
||||
// let output_path = config.output.path.clone();
|
||||
// let files = tables
|
||||
// .into_iter()
|
||||
// .map(|table| {
|
||||
// let output_path = output_path.clone();
|
||||
// let config = config.clone();
|
||||
// async move {
|
||||
// let table_name = match table.get_table_name() {
|
||||
// Some(table_ref) => match table_ref {
|
||||
// sea_query::TableRef::Table(t)
|
||||
// | sea_query::TableRef::SchemaTable(_, t)
|
||||
// | sea_query::TableRef::DatabaseSchemaTable(_, _, t)
|
||||
// | sea_query::TableRef::TableAlias(t, _)
|
||||
// | sea_query::TableRef::SchemaTableAlias(_, t, _)
|
||||
// | sea_query::TableRef::DatabaseSchemaTableAlias(_, _, t, _) => {
|
||||
// t.to_string()
|
||||
// }
|
||||
// _ => unimplemented!(),
|
||||
// },
|
||||
// None => return Err(eyre!("Table name not found")),
|
||||
// };
|
||||
// let table_str = generate_table(table, config.clone()).await?;
|
||||
// tracing::debug!(?table_str);
|
||||
// let filename = format!("{}.rs", table_name);
|
||||
// let file_path = output_path.join(&filename);
|
||||
// let exists = file_path.exists();
|
||||
// let content = if exists {
|
||||
// let mut file_content = fs::read_to_string(file_path).await?;
|
||||
// let rules = comment_parser::get_syntax("rust").unwrap();
|
||||
// let comments = CommentParser::new(&file_content, rules);
|
||||
// let mut found = false;
|
||||
// for comment in comments {
|
||||
// if let Event::BlockComment(a, b) = comment {
|
||||
// tracing::debug!(?a, ?b);
|
||||
// if b.contains(HEADER) {
|
||||
// found = true;
|
||||
// file_content = file_content.replace(a, &table_str);
|
||||
// tracing::debug!("Found header");
|
||||
// break;
|
||||
// }
|
||||
// }
|
||||
// }
|
||||
// if found {
|
||||
// Ok::<String, Report>(file_content)
|
||||
// } else {
|
||||
// let merged_content = format!("{}\n\n{}", table_str, file_content);
|
||||
// Ok::<String, Report>(merged_content)
|
||||
// }
|
||||
// } else {
|
||||
// Ok::<String, Report>(table_str)
|
||||
// }?;
|
||||
//
|
||||
// Ok(OutputFile {
|
||||
// name: filename,
|
||||
// content,
|
||||
// })
|
||||
// }
|
||||
// })
|
||||
// .collect::<JoinSet<Result<OutputFile>>>();
|
||||
// let files = files
|
||||
// .join_all()
|
||||
// .await
|
||||
// .into_iter()
|
||||
// .collect::<Result<Vec<OutputFile>>>()?;
|
||||
// Ok(files)
|
||||
// }
|
||||
|
||||
// async fn generate_table(table: TableCreateStatement, config: Config) -> Result<String> {
|
||||
// let mut string = String::new();
|
||||
// string.push_str(format!("{COMMENTHEAD} {HEADER}\n").as_str());
|
||||
//
|
||||
// let table_name = match table.get_table_name() {
|
||||
// Some(table_ref) => match table_ref {
|
||||
// sea_query::TableRef::Table(t)
|
||||
// | sea_query::TableRef::SchemaTable(_, t)
|
||||
// | sea_query::TableRef::DatabaseSchemaTable(_, _, t)
|
||||
// | sea_query::TableRef::TableAlias(t, _)
|
||||
// | sea_query::TableRef::SchemaTableAlias(_, t, _)
|
||||
// | sea_query::TableRef::DatabaseSchemaTableAlias(_, _, t, _) => t.to_string(),
|
||||
// _ => unimplemented!(),
|
||||
// },
|
||||
// None => return Err(eyre!("Table name not found")),
|
||||
// };
|
||||
// let mut inner = String::new();
|
||||
// inner.push_str(format!("{COMMENTBODY}\n Table name: {table_name}\n\n\n").as_str());
|
||||
// inner = inner.strip_suffix('\n').unwrap().to_string();
|
||||
// let mut ptable = Table::new();
|
||||
// let format = format::FormatBuilder::new()
|
||||
// .column_separator(' ')
|
||||
// .borders(' ')
|
||||
// .separators(
|
||||
// &[
|
||||
// format::LinePosition::Bottom,
|
||||
// format::LinePosition::Title,
|
||||
// // format::LinePosition::Top,
|
||||
// ],
|
||||
// format::LineSeparator::default(),
|
||||
// )
|
||||
// .padding(1, 1)
|
||||
// .build();
|
||||
// ptable.set_format(format);
|
||||
// ptable.set_titles(row!["Name", "Type", "RustType", "Attributes"]);
|
||||
// let indexes = table.get_indexes();
|
||||
// tracing::info!(?indexes);
|
||||
// for column in table.get_columns() {
|
||||
// let name = column.get_column_name();
|
||||
// if let Some(column_type) = column.get_column_type() {
|
||||
// let column_type_rust =
|
||||
// type_to_rust_string(column_type, config.sea_orm.entity.date_time_crate.clone());
|
||||
// let column_type =
|
||||
// type_to_string(column_type, config.sea_orm.entity.date_time_crate.clone());
|
||||
// let attrs = attrs_to_string(column.get_column_spec());
|
||||
// ptable.add_row(row![name, column_type, column_type_rust, attrs]);
|
||||
// }
|
||||
// }
|
||||
// inner.push_str(ptable.to_string().as_str());
|
||||
//
|
||||
// string.push_str(
|
||||
// inner
|
||||
// .replace("\n", format!("\n{} ", COMMENTBODY).as_str())
|
||||
// .as_str(),
|
||||
// );
|
||||
// string.push_str(format!("\n{COMMENTTAIL}\n").as_str());
|
||||
// Ok(string)
|
||||
// }
|
||||
//
|
||||
// fn type_to_string(column: &ColumnType, date_time_create: DateTimeCrate) -> String {
|
||||
// tracing::debug!(?column, ?date_time_create);
|
||||
// match column {
|
||||
// ColumnType::Char(Some(max)) => format!("Char({}max)", max),
|
||||
// ColumnType::Char(None) => "Char".to_owned(),
|
||||
// ColumnType::String(StringLen::None) => "String".to_owned(),
|
||||
// ColumnType::String(StringLen::Max) => "String(Max)".to_owned(),
|
||||
// ColumnType::String(StringLen::N(len)) => format!("String({}max)", len),
|
||||
// ColumnType::Text => "Text".to_owned(),
|
||||
// ColumnType::Integer => "Integer".to_owned(),
|
||||
//
|
||||
// ColumnType::Custom(v) => v.to_string(),
|
||||
// _ => unimplemented!(),
|
||||
// }
|
||||
// }
|
||||
// fn type_to_rust_string(column: &ColumnType, date_time_create: DateTimeCrate) -> String {
|
||||
// tracing::debug!(?column, ?date_time_create);
|
||||
// match column {
|
||||
// ColumnType::Char(_) | ColumnType::String(_) | ColumnType::Text | ColumnType::Custom(_) => {
|
||||
// "String".to_owned()
|
||||
// }
|
||||
// ColumnType::Integer => "i32".to_owned(),
|
||||
// _ => unimplemented!(),
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// fn attrs_to_string(column: &Vec<ColumnSpec>) -> String {
|
||||
// tracing::debug!(?column);
|
||||
// column
|
||||
// .iter()
|
||||
// .filter_map(|c| match c {
|
||||
// ColumnSpec::PrimaryKey => Some("primary key"),
|
||||
// ColumnSpec::UniqueKey => Some("unique key"),
|
||||
// ColumnSpec::AutoIncrement => Some("autoincrement"),
|
||||
// ColumnSpec::NotNull => Some("not null"),
|
||||
// _ => None,
|
||||
// })
|
||||
// .map(|s| s.to_string())
|
||||
// .collect::<Vec<String>>()
|
||||
// .join(", ")
|
||||
// }
|
||||
//
|
||||
// //
|
||||
46
src/generator/table.rs
Normal file
46
src/generator/table.rs
Normal file
@@ -0,0 +1,46 @@
|
||||
use super::column::Column;
|
||||
use crate::config::Config;
|
||||
use color_eyre::{eyre::eyre, Result};
|
||||
use sea_schema::sea_query::{self, ColumnDef, IndexCreateStatement, TableCreateStatement};
|
||||
use tracing::instrument;
|
||||
|
||||
#[derive(Debug, Clone)]
|
||||
pub struct Table {
|
||||
pub name: String,
|
||||
pub columns: Vec<Column>,
|
||||
}
|
||||
|
||||
impl Table {
|
||||
pub fn new(statement: TableCreateStatement, config: Config) -> Result<Table> {
|
||||
let table_name = match statement.get_table_name() {
|
||||
Some(table_ref) => match table_ref {
|
||||
sea_query::TableRef::Table(t)
|
||||
| sea_query::TableRef::SchemaTable(_, t)
|
||||
| sea_query::TableRef::DatabaseSchemaTable(_, _, t)
|
||||
| sea_query::TableRef::TableAlias(t, _)
|
||||
| sea_query::TableRef::SchemaTableAlias(_, t, _)
|
||||
| sea_query::TableRef::DatabaseSchemaTableAlias(_, _, t, _) => t.to_string(),
|
||||
_ => unimplemented!(),
|
||||
},
|
||||
None => return Err(eyre!("Table name not found")),
|
||||
};
|
||||
tracing::debug!(?table_name);
|
||||
let columns_raw = statement.get_columns();
|
||||
let indexes = statement.get_indexes();
|
||||
let columns = columns_raw
|
||||
.iter()
|
||||
.map(|column| {
|
||||
let name = column.get_column_name();
|
||||
let index = indexes
|
||||
.iter()
|
||||
.find(|index| index.get_index_spec().get_column_names().contains(&name));
|
||||
Column::new(column.clone(), index.cloned())
|
||||
})
|
||||
.collect::<Result<Vec<Column>>>()?;
|
||||
tracing::debug!(?columns);
|
||||
Ok(Table {
|
||||
columns,
|
||||
name: table_name,
|
||||
})
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user