Cargo update & some doc tweak
This commit is contained in:
parent
aad12d829f
commit
d524482ac1
406
Cargo.lock
generated
406
Cargo.lock
generated
File diff suppressed because it is too large
Load diff
|
@ -5,7 +5,7 @@ authors = ["Vincent Prouillet <prouillet.vincent@gmail.com>"]
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
tera = "0.11"
|
tera = "0.11"
|
||||||
base64 = "0.9"
|
base64 = "0.10"
|
||||||
lazy_static = "1"
|
lazy_static = "1"
|
||||||
pulldown-cmark = "0"
|
pulldown-cmark = "0"
|
||||||
toml = "0.4"
|
toml = "0.4"
|
||||||
|
|
|
@ -112,7 +112,7 @@ impl Hash for DataSource {
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
fn get_data_from_args(content_path: &PathBuf, args: &HashMap<String, Value>) -> Result<DataSource> {
|
fn get_data_source_from_args(content_path: &PathBuf, args: &HashMap<String, Value>) -> Result<DataSource> {
|
||||||
let path_arg = optional_arg!(
|
let path_arg = optional_arg!(
|
||||||
String,
|
String,
|
||||||
args.get("path"),
|
args.get("path"),
|
||||||
|
@ -157,15 +157,15 @@ fn get_output_format_from_args(args: &HashMap<String, Value>, data_source: &Data
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
/// A global function to load data from a data file.
|
/// A global function to load data from a file or from a URL
|
||||||
/// Currently the supported formats are json, toml and csv
|
/// Currently the supported formats are json, toml, csv and plain text
|
||||||
pub fn make_load_data(content_path: PathBuf, base_path: PathBuf) -> GlobalFn {
|
pub fn make_load_data(content_path: PathBuf, base_path: PathBuf) -> GlobalFn {
|
||||||
let mut headers = header::HeaderMap::new();
|
let mut headers = header::HeaderMap::new();
|
||||||
headers.insert(header::USER_AGENT, "zola".parse().unwrap());
|
headers.insert(header::USER_AGENT, "zola".parse().unwrap());
|
||||||
let client = Arc::new(Mutex::new(Client::builder().build().expect("reqwest client build")));
|
let client = Arc::new(Mutex::new(Client::builder().build().expect("reqwest client build")));
|
||||||
let result_cache: Arc<Mutex<HashMap<u64, Value>>> = Arc::new(Mutex::new(HashMap::new()));
|
let result_cache: Arc<Mutex<HashMap<u64, Value>>> = Arc::new(Mutex::new(HashMap::new()));
|
||||||
Box::new(move |args| -> Result<Value> {
|
Box::new(move |args| -> Result<Value> {
|
||||||
let data_source = get_data_from_args(&content_path, &args)?;
|
let data_source = get_data_source_from_args(&content_path, &args)?;
|
||||||
|
|
||||||
let file_format = get_output_format_from_args(&args, &data_source)?;
|
let file_format = get_output_format_from_args(&args, &data_source)?;
|
||||||
|
|
||||||
|
@ -200,23 +200,19 @@ pub fn make_load_data(content_path: PathBuf, base_path: PathBuf) -> GlobalFn {
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
/// load/parse a json file from the given path and place it into a
|
/// Parse a JSON string and convert it to a Tera Value
|
||||||
/// tera value
|
|
||||||
fn load_json(json_data: String) -> Result<Value> {
|
fn load_json(json_data: String) -> Result<Value> {
|
||||||
let json_content: Value = serde_json::from_str(json_data.as_str()).map_err(|e| format!("{:?}", e))?;
|
let json_content: Value = serde_json::from_str(json_data.as_str()).map_err(|e| format!("{:?}", e))?;
|
||||||
return Ok(json_content);
|
return Ok(json_content);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// load/parse a toml file from the given path, and place it into a
|
/// Parse a TOML string and convert it to a Tera Value
|
||||||
/// tera Value
|
|
||||||
fn load_toml(toml_data: String) -> Result<Value> {
|
fn load_toml(toml_data: String) -> Result<Value> {
|
||||||
let toml_content: toml::Value = toml::from_str(&toml_data).map_err(|e| format!("{:?}", e))?;
|
let toml_content: toml::Value = toml::from_str(&toml_data).map_err(|e| format!("{:?}", e))?;
|
||||||
|
|
||||||
to_value(toml_content).map_err(|e| e.into())
|
to_value(toml_content).map_err(|e| e.into())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Load/parse a csv file from the given path, and place it into a
|
/// Parse a CSV string and convert it to a Tera Value
|
||||||
/// tera Value.
|
|
||||||
///
|
///
|
||||||
/// An example csv file `example.csv` could be:
|
/// An example csv file `example.csv` could be:
|
||||||
/// ```csv
|
/// ```csv
|
||||||
|
@ -236,7 +232,6 @@ fn load_toml(toml_data: String) -> Result<Value> {
|
||||||
/// ```
|
/// ```
|
||||||
fn load_csv(csv_data: String) -> Result<Value> {
|
fn load_csv(csv_data: String) -> Result<Value> {
|
||||||
let mut reader = Reader::from_reader(csv_data.as_bytes());
|
let mut reader = Reader::from_reader(csv_data.as_bytes());
|
||||||
|
|
||||||
let mut csv_map = Map::new();
|
let mut csv_map = Map::new();
|
||||||
|
|
||||||
{
|
{
|
||||||
|
@ -257,7 +252,6 @@ fn load_csv(csv_data: String) -> Result<Value> {
|
||||||
|
|
||||||
for result in records {
|
for result in records {
|
||||||
let record = result.unwrap();
|
let record = result.unwrap();
|
||||||
|
|
||||||
let mut elements_array: Vec<Value> = Vec::new();
|
let mut elements_array: Vec<Value> = Vec::new();
|
||||||
|
|
||||||
for e in record.into_iter() {
|
for e in record.into_iter() {
|
||||||
|
|
Loading…
Reference in a new issue