diff options
Diffstat (limited to 'web/vokobe/src/main.rs')
-rw-r--r-- | web/vokobe/src/main.rs | 922 |
1 files changed, 922 insertions, 0 deletions
diff --git a/web/vokobe/src/main.rs b/web/vokobe/src/main.rs new file mode 100644 index 0000000..ab26457 --- /dev/null +++ b/web/vokobe/src/main.rs @@ -0,0 +1,922 @@ +/* +pull the std into scope and inline it so that we get documentation for it, +even when running offline +*/ +#[doc(inline)] +pub use std; + +use std::path::{Path, PathBuf}; +use std::io::{self, Read, Write, BufRead, BufReader}; +use std::fs::{self, File}; +use std::time; +use std::collections::HashMap; + +use structopt::StructOpt; +use regex::Regex; + +#[derive(Debug, StructOpt)] +#[structopt(name = "vokobe", about = "A static site generator")] +struct Opt { + /// Input path + #[structopt(parse(from_os_str))] + input_path: PathBuf, + + /// Output path + #[structopt(parse(from_os_str))] + output_path: PathBuf, + + /// Site name (e.g. emile.space) + site_name: String, + + /// Activate sending analytics to stats.emile.space + // -a and --analytics will be generated + // analytics are sent to stats.emile.space + #[structopt(short, long)] + analytics: bool, +} + +fn main() -> std::io::Result<()> { + + let mut internal_links: HashMap<String, Vec<String>> = HashMap::new(); + + let opt = Opt::from_args(); + + let in_path = opt.input_path; + let output_path = opt.output_path; + + // read the style + let style_path = Path::new(&in_path).join("style.css"); + let mut style_file = File::open(style_path) + .expect("could not open style file"); + let mut style = String::new(); + style_file.read_to_string(&mut style) + .expect("could not read style file to string"); + + // read all dirs in the input path + let pathes = recursive_read_dir(&in_path, false)?; + + // pass 1: store the backlinks + + for path in &pathes { + if path.ends_with("README.md") { + // open the file and read it as a string + let mut readme_file = File::open(path)?; + let mut readme = String::new(); + readme_file.read_to_string(&mut readme)?; + + let internal_links_in_file + = parse_internal_links(readme.as_str()); + + for link in internal_links_in_file { + + internal_links.entry(link).or_insert_with(Vec::new).push(path.to_string_lossy().into_owned()) + } + } + } + + + // for each markdown_file in markdown_files { + // let internal_links_in_file = parse_internal_links(markdown_file); + // internal_links.insert(markdown_file, internal_links_in_file); + // } + + // pass 2: create the html + + println!("Got {} files", pathes.len()); + let mut readme_counter = 0; + + for path in pathes { + let stripped_path = path.strip_prefix(&in_path) + .expect(format!( + "could not strip the in_path prefix: {:?}", in_path).as_str()); + + // copy images and other files to the output folder + if path.is_file() { + + // define the source and destination + let src = Path::new(&in_path).join(stripped_path); + let dst = Path::new(&output_path).join(stripped_path); + + // define the destination folder (the dst path without the file) and create it + let mut dst_folder = dst.clone(); + dst_folder.pop(); // remove the file itself from the path + fs::create_dir_all(dst_folder)?; + + // copy the file to the destination + std::fs::copy(src, dst.as_path())?; + } + + if stripped_path.ends_with("README.md") { + readme_counter += 1; + + // define the "raw" path (no infile prefix, no file) + let mut ancestors = stripped_path.ancestors(); + ancestors.next(); + + let raw_path = ancestors.next() + .expect("could not extract next ancestor"); + + // out + rawpath + let index_path = output_path.join(raw_path); + + // (out + rawpath) + "index.html" + let index_file = index_path.join("index.html"); + + // - create the dir for the index.html as well as the index.html + // itself + fs::create_dir_all(index_path)?; + let mut file = File::create(&index_file)?; + + // this is the main block calling all other smaller functions. The + // whole output is compsed here + write_header(&mut file, &opt.site_name, &style)?; + write_body_start(&mut file, &opt.site_name)?; + write_nav(&mut file, in_path.as_path(), raw_path, opt.analytics)?; + write_same_level(&mut file, in_path.as_path(), raw_path)?; + write_readme_content(&mut file, in_path.as_path(), raw_path)?; + write_footer(&mut file, raw_path, &internal_links)?; + + file.write_all("".as_bytes())?; + } + } + + println!("Got {readme_counter} README.md files"); + + Ok(()) +} + +fn parse_internal_links(markdown_file: &str) -> Vec<String> { + // Define a regular expression to match markdown-style links + let link_regex = Regex::new(r"\[([^\]]+)\]\(([^)]+)\)").unwrap(); + + // Initialize a vector to store internal links found in the markdown file + let mut internal_links = Vec::new(); + + // Iterate over each match of the regular expression in the markdown content + for capture in link_regex.captures_iter(&markdown_file) { + // Extract the link text and URL from the capture groups + // let link_text = &capture[1]; + let mut link_url = &capture[2]; + + // Check if the link is an internal link (e.g., relative URL) + // You can customize this condition based on your site's URL structure + if link_url.starts_with('/') || link_url.starts_with("../") { + if link_url.ends_with('/') { + link_url = link_url.trim_end_matches('/'); + } + internal_links.push(link_url.to_string()); + } + } + + internal_links +} + +/// Write the html header including the style file +/// TODO: Don't add the style file into each compiled html output, as the +/// style can be included allowing the user to cache the style file in their +/// browser. +fn write_header(file: &mut File, site_name: &String, style: &String) -> std::io::Result<()>{ + + // write the header including the style file + file.write_all(format!(r#"<!DOCTYPE html> +<html lang="en"> +<head> + <meta charset="UTF-8"> + <meta http-equiv="X-UA-Compatible" content="IE=edge"> + <meta name="viewport" content="width=device-width, initial-scale=1.0"> + <title>{}</title> + + <style> + {} + </style> +</head> + "#, site_name, style).as_bytes())?; + + Ok(()) +} + +/// write the start of the html body tag and the header linking back to the +/// site itself. +fn write_body_start(file: &mut File, site_name: &String) -> std::io::Result<()>{ + file.write_all(format!(r#" +<body> + <header> + <a href="/">{}</a> + </header>"#, site_name).as_bytes())?; + + Ok(()) +} + +/// Write the navigation section to the given file +fn write_nav(file: &mut File, in_path: &Path, raw_path: &Path, analytics: bool) + -> std::io::Result<()> { + + if analytics == true { + /* + file.write_all(format!(r#" + <img src="https://stats.emile.space/count?p=/{}"> + <nav> + <ul>"#, raw_path.to_str().unwrap()).as_bytes())?; + */ + file.write_all(format!(r#" + <nav> + <ul>"#,).as_bytes())?; + } else { + file.write_all(format!(r#" + <nav> + <ul>"#).as_bytes())?; + } + + // get the nav bar components + let components = raw_path.components().collect::<Vec<_>>(); + + // for each list of components (["a"], ["a", "b"], ["a", "b", "c"]), create + // the path for the list, view all other dirs at that path and write the + // result to the file + let mut i = 0; + let slice = components.as_slice(); + + // for each navbar component + for component in slice { + + // get the items belonging to that navbar item + // (["a"], ["a", "b"], ["a", "b", "c"]) + let subpath_components = &slice[..i+1]; + i += 1; + + let mut subpath_path = PathBuf::new(); + + // push the inpath, so we've got a basis from where we can read the + // subpath items + // subpath_path = inpath + ??? + subpath_path.push(in_path); + + let mut nav_breadcrumb_link = PathBuf::new(); + + // for each item in the subpath, push it into the subpath_path so that + // in the end, we've got something like this: + // "inpath" + "a" + "b" + "c" + for subpath_component in subpath_components { + subpath_path.push(subpath_component); + nav_breadcrumb_link.push(subpath_component); + } + + // make the nav_breadcrumb_link an absolute by prefixing it with a / + // (this is in scope of the web-page, so this is find) and make it a + // string + let nav_breadcrumb_link_absolute + = Path::new("/") + .join(nav_breadcrumb_link); + + let nav_breadcrumb_link + = nav_breadcrumb_link_absolute.to_str().unwrap(); + + // define the name of the breadcrumb + let nav_breadcrumb_name = component.as_os_str().to_str().unwrap(); + + //////////////////////////////////////////////////////////////////////// + file.write_all(format!(r#" + <li> + <a href="{}">{}</a> + <ul>"#, nav_breadcrumb_link, nav_breadcrumb_name).as_bytes())?; + //////////////////////////////////////////////////////////////////////// + + // as we don't want to get the items for the individial entry, but on + // the same level, we push a ".." + // the subpath_path is now: inpath + subpath + ../ + subpath_path.push(".."); + + // read all dirs in the subpath_path, add them to the dirs vector, so + // that we get a vector containing all the dirs we want + let mut dirs = Vec::new(); + for entry in fs::read_dir(subpath_path)? { + let path = &entry?.path(); + if path.is_dir() { + dirs.push(path.to_path_buf()); + } + } + + dirs.sort(); + + // DROPDOWN + // extract the link and name for each directory found + for dir in dirs { + let d = dir.canonicalize()?; + let abs_inpath = in_path.canonicalize()?; + + let name = d.file_name().unwrap().to_str().unwrap(); + let rel_link + = d.strip_prefix(abs_inpath) + .expect(format!( + "could not strip the in_path prefix: {:?}", + d).as_str()); + + let link = Path::new("/").join(rel_link); + let link = link.as_path().to_str().unwrap(); + + // don't add the current page to the dropdown, we're on it already! + if name == nav_breadcrumb_name { + continue + } + + // don't add items starting with a dot to the dropdown, they're + // hidden! + if name.starts_with(".") { + continue + } + + //////////////////////////////////////////////////////////////////// + file.write_all(format!(r#" + <li><a href="{}">{}/</a></li>"#, link, name).as_bytes())?; + //////////////////////////////////////////////////////////////////// + } + + //////////////////////////////////////////////////////////////////////// + file.write_all(r#" + </ul> + </li>"#.as_bytes())?; + //////////////////////////////////////////////////////////////////////// + } + + //////////////////////////////////////////////////////////////////////////// + file.write_all(format!(r#" + </ul> + <ul style="float: right"> + <li>{:?}</li> + <li> + <a href="README.md">.md</a> + </li> + </ul> + </nav>"#, in_path.metadata()?.modified()?.duration_since(std::time::UNIX_EPOCH).unwrap().as_secs()).as_bytes())?; + //////////////////////////////////////////////////////////////////////////// + + Ok(()) +} + + +fn write_same_level(file: &mut File, in_path: &Path, raw_path: &Path) + -> std::io::Result<()> { + + let search_path = Path::new(in_path).join(raw_path); + + let mut dirs: Vec<PathBuf> = Vec::new(); + let mut files: Vec<PathBuf> = Vec::new(); + + let mut vertical: bool = false; + let mut show_files: bool = false; + + for entry in fs::read_dir(search_path)? { + let path = &entry?.path(); + + if path.is_dir() { + dirs.push(path.to_path_buf()); + } + if path.is_file() { + files.push(path.to_path_buf()); + if path.file_name().unwrap() == "vertical" { + vertical = true; + } + if path.file_name().unwrap() == "show_files" { + show_files = true; + } + } + } + + dirs.sort(); + files.sort(); + + let in_path = in_path.canonicalize()?; + + if vertical == true { + file.write_all(format!(r#" + <ul class="vert">"#).as_bytes())?; + } else { + file.write_all(format!(r#" + <ul>"#).as_bytes())?; + } + + for dir in dirs { + let dir = dir.canonicalize()?; + let dir = dir.strip_prefix(&in_path) + .expect("could not strip in_path prefix"); + + let link = Path::new("/").join(dir); + let link_str = link.as_path().to_str().unwrap(); + let name = link.file_name().unwrap().to_str().unwrap(); + + if name.starts_with(".") { + continue + } + + file.write_all(format!(r#" + <li><a href="{}">{}/</a></li>"#, link_str, name).as_bytes())?; + } + + file.write_all(format!(r#" + </ul>"#).as_bytes())?; + + if files.len() >= 1 && show_files == true { + file.write_all(format!(r#"<br> + <ul>"#).as_bytes())?; + + for f in files { + let f = f.canonicalize()?; + let f = f.strip_prefix(&in_path) + .expect("could not strip in_path prefix"); + + let link = Path::new("/").join(f); + let link_str = link.as_path().to_str().unwrap(); + let name = link.file_name().unwrap().to_str().unwrap(); + + if name == "README.md" + || name == "show_files" + || name.starts_with(".") + { + continue + }; + + file.write_all(format!(r#" + <li><a href="{}">{}</a></li>"#, link_str, name).as_bytes())?; + } + + file.write_all(format!(r#" + </ul>"#).as_bytes())?; + } + + + Ok(()) +} + +fn write_readme_content(file: &mut File, in_path: &Path, raw_path: &Path) + -> std::io::Result<()> { + + // define the path of the README.md file + let readme_file_path + = Path::new(in_path).join(raw_path).join("README.md"); + + // open the file and read it as a string + let mut readme_file = File::open(readme_file_path)?; + let mut readme = String::new(); + readme_file.read_to_string(&mut readme)?; + + // replace all "markdown" style links with HTML links + // let re = Regex::new(r"\[([^\[]+)\]\(([^\(]+)\)").unwrap(); + let re = Regex::new(r"\[([^]]+)\]\(([^)]+)\)").unwrap(); + let readme = re.replace_all(&readme, "<a href=\"$2\">$1</a>"); + + file.write_all(format!("<pre>").as_bytes())?; + + // counting the occurrence of `---` + let mut hrule_count = 0; + let mut in_yaml_metadata_block= false; + + let mut level_1_heading_num = 0; + let mut level_2_heading_num = 0; + let mut level_3_heading_num = 0; + let mut level_4_heading_num = 0; + let mut level_5_heading_num = 0; + + // cheap markdown 2 html converter + for line in readme.split('\n') { + + // 1 == 2, as I'm not sure how to comment out the file write 5 lines or so below + if in_yaml_metadata_block && 1 == 2 { + // if we find the end of the yaml metadata block, break this + if line.starts_with("---") { + in_yaml_metadata_block = false; + continue + } else { + file.write_all(format!(r##"yaml_line: {} +"##, line).as_bytes())?; + continue + } + } + + // if we've got a horizontal rule, it can be two things: the start and + // end of a yaml-metadata block or an actual horizontal rule. + // + // If it's yaml metadata, read it all, but don't print it, store it + // for later + // If it's a horizontal rule, print the horizontal rule + if line.starts_with("---") { + + // store the yaml metadata + if hrule_count == 0 { + in_yaml_metadata_block = true; + continue + } + hrule_count += 1; + + // print the horizontal rule + file.write_all(format!(r##" + <hr>"##).as_bytes())?; + + } else if line.starts_with("#####") { + let heading = line.get(6..).unwrap(); + let heading_sanitized = sanitize(heading.to_string()); + + level_5_heading_num += 1; + + file.write_all(format!(r##"</pre> + <span id="{a}"></span> + <h5><a href="#{a}">{h1}.{h2}.{h3}.{h4}.{h5}. {b}</a></h3> + <pre>"##, + a = heading_sanitized, + b = heading, + h1 = level_1_heading_num, + h2 = level_2_heading_num, + h3 = level_3_heading_num, + h4 = level_4_heading_num, + h5 = level_5_heading_num, + ).as_bytes())?; + + } else if line.starts_with("####") { + let heading = line.get(5..).unwrap(); + let heading_sanitized = sanitize(heading.to_string()); + + level_4_heading_num += 1; + level_5_heading_num = 0; + + file.write_all(format!(r##"</pre> + <span id="{a}"></span> + <h4><a href="#{a}">{h1}.{h2}.{h3}.{h4}. {b}</a></h3> + <pre>"##, + a = heading_sanitized, + b = heading, + h1 = level_1_heading_num, + h2 = level_2_heading_num, + h3 = level_3_heading_num, + h4 = level_4_heading_num, + ).as_bytes())?; + + } else if line.starts_with("###") { + let heading = line.get(4..).unwrap(); + let heading_sanitized = sanitize(heading.to_string()); + + level_3_heading_num += 1; + level_4_heading_num = 0; + level_5_heading_num = 0; + + file.write_all(format!(r##"</pre> + <span id="{a}"></span> + <h3><a href="#{a}">{h1}.{h2}.{h3}. {b}</a></h3> + <pre>"##, + a = heading_sanitized, + b = heading, + h1 = level_1_heading_num, + h2 = level_2_heading_num, + h3 = level_3_heading_num, + ).as_bytes())?; + + } else if line.starts_with("##") { + let heading = line.get(3..).unwrap(); + let heading_sanitized = sanitize(heading.to_string()); + + level_2_heading_num += 1; + level_3_heading_num = 0; + level_4_heading_num = 0; + level_5_heading_num = 0; + + file.write_all(format!(r##"</pre> + <span id="{a}"></span> + <h2><a href="#{a}">{h1}.{h2}. {b}</a></h2> + <pre>"##, + a = heading_sanitized, + b = heading, + h1 = level_1_heading_num, + h2 = level_2_heading_num, + ).as_bytes())?; + + } else if line.starts_with("#") { + let heading = line.get(2..).unwrap(); + let heading_sanitized = sanitize(heading.to_string()); + + level_1_heading_num += 1; + level_2_heading_num = 0; + level_3_heading_num = 0; + level_4_heading_num = 0; + level_5_heading_num = 0; + + file.write_all(format!(r##"</pre> + <span id="{a}"></span> + <h1><a href="#{a}">{h1}. {b}</a></h1> + <pre>"##, + a = heading_sanitized, + b = heading, + h1 = level_1_heading_num + ).as_bytes())?; + + } else if line.starts_with("> ") { + let line = line.replace("<", "<"); + let line = line.get(2..).unwrap(); + file.write_all(format!("</pre><pre class=\"code\">{}</pre><pre>\n", line).as_bytes())?; + + } else if line.starts_with(":::tree") { + + // TODO: add some parameter controlling if the list is ascending or descending (reverse the list before writing) + + // get all dirs in the current dir recursively + let tree_files_path = Path::new(in_path).join(raw_path); + let mut tree_files + = recursive_read_dir(&tree_files_path, true)?; + + // sort them, otherwise we'll get complete chaos + tree_files.sort(); + + for path in tree_files { + + // strip the inpath prefix and raw_path prefix, as we don't need + // them + let path + = path.strip_prefix(in_path) + .expect("could not strip in_file prefix") + .strip_prefix(raw_path) + .expect("could not strip raw_path prefix"); + + // convert the path to a string, check if it contains a hidden + // path by checking if it contains a `/.`, if so, skip this one + if String::from(path.to_str().unwrap()).contains("/.") { + continue + } + if String::from(path.to_str().unwrap()).starts_with(".") { + continue + } + + // write the link and the entry name to the file + let link = Path::new(raw_path).join(path); + let name = path.file_name().unwrap().to_str().unwrap(); + + // count the amount of segments in the path and write spaces for + // each + let segments = path.iter().count(); + for _ in 0..(segments-1) { + file.write_all(r#" "#.as_bytes())?; + } + + file.write_all( + format!("<a href=\"/{}\">{}</a>\n", + link.display(), name, + ).as_bytes() + )?; + } + + } else if line.starts_with(":::toc") { + + // TODO: depth parameter for controlling the depth of the table of contents + + let mut level_1_num = 0; + let mut level_2_num = 0; + let mut level_3_num = 0; + let mut level_4_num = 0; + let mut level_5_num = 0; + + for line in readme.split('\n') { + if line.starts_with("#####") { + let line = line.get(6..).unwrap(); + // trim the line to remove the trailing whitespace + let line = line.trim(); + level_5_num += 1; + file.write_all( + format!( + r##" <a href="#{}">{}.{}.{}.{}.{}. {}</a> +"##, + sanitize(line.to_string()), + level_1_num, + level_2_num, + level_3_num, + level_4_num, + level_5_num, + line + ).as_bytes() + )?; + } else if line.starts_with("####") { + let line = line.get(5..).unwrap(); + // trim the line to remove the trailing whitespace + let line = line.trim(); + level_4_num += 1; + level_5_num = 0; + file.write_all( + format!( + r##" <a href="#{}">{}.{}.{}.{}. {}</a> +"##, + sanitize(line.to_string()), + level_1_num, + level_2_num, + level_3_num, + level_4_num, + line + ).as_bytes() + )?; + } else if line.starts_with("###") { + let line = line.get(4..).unwrap(); + // trim the line to remove the trailing whitespace + let line = line.trim(); + level_3_num += 1; + level_4_num = 0; + level_5_num = 0; + file.write_all( + format!( + r##" <a href="#{}">{}.{}.{}. {}</a> +"##, + sanitize(line.to_string()), + level_1_num, + level_2_num, + level_3_num, + line + ).as_bytes() + )?; + } else if line.starts_with("##") { + let line = line.get(3..).unwrap(); + let line = line.trim(); + level_2_num += 1; + level_3_num = 0; + level_4_num = 0; + level_5_num = 0; + + file.write_all( + format!( + //r##" <a href="#{}">{}.{}. {}</a> + r##" <a href="#{}">{}.{}. {}</a> +"##, + sanitize(line.to_string()), + level_1_num, + level_2_num, + line + ).as_bytes() + )?; + } else if line.starts_with("#") { + let line = line.get(2..).unwrap(); + let line = line.trim(); + level_1_num += 1; + level_2_num = 0; + level_3_num = 0; + level_4_num = 0; + level_5_num = 0; + + file.write_all( + format!( + r##"<a href="#{}">{}. {}</a> +"##, + sanitize(line.to_string()), + level_1_num, + line + ).as_bytes() + )?; + } + } + + } else { + + // for the case that nothing of the above matches, just write the + // content into the html body as it is + file.write_all(format!("{}\n", line).as_bytes())?; + } + } + + Ok(()) +} + +fn write_footer(file: &mut File, raw_path: &Path, internal_links: &HashMap<String, Vec<String>>) -> std::io::Result<()> { + + // add some padding before the whole footer stuff + file.write_all(b"<br><br><br>")?; + + // Backlinks + + let search_path = Path::new("/").join(raw_path).into_os_string().into_string().unwrap(); + + match internal_links.get(&search_path) { + Some(values) => { + + // only write "backlinks" if we've actually got some + file.write_all(b"backlinks:\n")?; + + for link in values { + + // strip the "in" prefix + // strip the "README.md" suffix + // TODO: do all this magic by parsing it as a path and removing the unneeded parts, bonus by creating a function doing this and removing the horrible string mashing in this codebase + let a = link + .strip_prefix("in") + .expect("no prefix to strip") + .strip_suffix("README.md") + .expect("no README.md suffix to remove"); + + file.write_all(format!(r#"- <a href="{a}">{a}</a> +"#).as_bytes())?; + } + } + None => (), + } + + // The actual footer + + file.write_all(format!(r#" + </pre> +<a href="https://chaos.social/@hanemile.rss" target="_blank" rel="noopener" class="icon"><img class="webring" src="/rss.svg" alt="rss feed of @hanemile@chaos.social mastodon" height="32px"/></a> +<a href="https://lieu.cblgh.org/" target="_blank" rel="noopener" class="icon"><img class="webring" src="/lieu.svg" alt="lieu webring search engine" height="32px"/></a> +<a href="https://webring.xxiivv.com/#emile" target="_blank" rel="noopener" class="icon"><img class="webring" src="/webring.svg" alt="XXIIVV webring" height="32px"/></a> +<a rel="me" href="https://chaos.social/@hanemile" target="_blank" class="icon"><img class="webring" src="/mastodon.svg" alt="mastodon" height="32px"/></a> + <pre>emile - {:?} - generated using <a href="https://github.com/hanemile/vokobe">vokobe {:?}</a><pre> +</body> +</html> +"#, + time::SystemTime::now().duration_since(time::SystemTime::UNIX_EPOCH).unwrap(), + env!("CARGO_PKG_VERSION") + ).as_bytes())?; + + Ok(()) +} + +/// sanitize the given string (to lower + space to hypen + keep only +/// [a-zA-Z0-9]) +fn sanitize(input: String) -> String { + let input = input.replace(" ", "-"); + + input + .chars() + .filter(|c| c.is_ascii_alphanumeric() || c.eq(&'-')) + .collect::<String>() + .to_lowercase() +} + +/// Return a list of all files in the directory, recursively. +fn recursive_read_dir(dir: &PathBuf, dir_only: bool) -> io::Result<Vec<PathBuf>> { + + // return an empty vec if the given path is not a directory + if dir.is_dir() == false { + return Ok(vec![]); + } + + if dir.starts_with(".") { + return Ok(vec![]); + } + + // get all entries in the gitignore file, if it exists + let gitignore_entries: Vec<PathBuf> = gitignore_entries(&dir)?; + + // store the child pathes + let mut entries: Vec<PathBuf> = Vec::new(); + + // iterate over all items in the dir, pushing the dirs pathes to the dirs + // vector for returning it + 'outer: for entry in fs::read_dir(dir)? { + let dir_entry = &entry?; + let path = dir_entry.path(); + + // skip hidden folders + if path.starts_with(".") { + //continue 'outer; + break 'outer; + } + if dir.starts_with(".") { + //continue 'outer; + break 'outer; + } + + // check if the current entry is part of the gitignore, if so, skip it + for gitignore_entry in &gitignore_entries { + if gitignore_entry.to_str() == Some("") { + continue; + } + if path.ends_with(gitignore_entry) { + continue 'outer; + } + } + + if dir_only == true { + if path.is_dir() { + entries.push(path.to_path_buf()); + } + } else { + entries.push(path.to_path_buf()); + } + + // recursively push all dirs from all children to the dirs vector + let subdirs = recursive_read_dir(&path, dir_only)?; + + for subdir in subdirs { + entries.push(subdir) + } + } + + // return the dirs, the ones from this folder and the ones from all child folders + Ok(entries) +} + +// try to open the gitignore file and read all entries from there. +fn gitignore_entries(dir: &PathBuf) -> io::Result<Vec<PathBuf>> { + let gitignore_path = Path::new(&dir) + .join(Path::new(".gitignore")); + + let mut entries: Vec<PathBuf> = Vec::new(); + if let Ok(gitignore) = File::open(&gitignore_path) { + let reader = BufReader::new(gitignore); + + for line in reader.lines() { + entries.push(PathBuf::from(line?)); + } + } + + Ok(entries) +} |