Compare commits
2 commits
72e797e6e6
...
dfbb3577ed
Author | SHA1 | Date | |
---|---|---|---|
Reynir Björnsson | dfbb3577ed | ||
Reynir Björnsson | a2722c0947 |
100
posts/gptar-update.md
Normal file
100
posts/gptar-update.md
Normal file
|
@ -0,0 +1,100 @@
|
||||||
|
---
|
||||||
|
title: GPTar (update)
|
||||||
|
date: 2024-10-28
|
||||||
|
description: libarchive vs hybrid GUID partition table and GNU tar volume header
|
||||||
|
---
|
||||||
|
|
||||||
|
In a [previous post][gptar-post] I describe how I craft a hybrid GUID partition table (GPT) and tar archive by exploiting that there are disjoint areas of a 512 byte *block* that are important to tar headers and *protective* master boot records used in GPT respectively.
|
||||||
|
I recommend reading it first if you haven't already for context.
|
||||||
|
|
||||||
|
After writing the above post I read an excellent and fun *and totally normal* article by Emily on how [she created **executable** tar archives][tar-executable].
|
||||||
|
Therein I learned a clever hack:
|
||||||
|
GNU tar has a tar extension for *volume headers*.
|
||||||
|
These are essentially labels for your tape archives when you're forced to split an archive across multiple tapes.
|
||||||
|
They can (seemingly) hold any text as label including shell scripts.
|
||||||
|
What's more is GNU tar and bsdtar **does not** extract these as files!
|
||||||
|
This is excellent, because I don't actually want to extract or list the GPT header when using GNU tar or bsdtar.
|
||||||
|
This prompted me to [use a different link indicator](https://github.com/reynir/gptar/pull/1).
|
||||||
|
|
||||||
|
This worked pretty great.
|
||||||
|
Listing the archive using GNU tar I still get `GPTAR`, but with verbose listing it's displayed as a `--Volume Header--`:
|
||||||
|
|
||||||
|
```
|
||||||
|
$ tar -tvf disk.img
|
||||||
|
Vr-------- 0/0 16896 1970-01-01 01:00 GPTAR--Volume Header--
|
||||||
|
-rw-r--r-- 0/0 14 1970-01-01 01:00 test.txt
|
||||||
|
```
|
||||||
|
|
||||||
|
And more importantly the `GPTAR` entry is ignored when extracting:
|
||||||
|
|
||||||
|
```
|
||||||
|
$ mkdir tmp
|
||||||
|
$ cd tmp/
|
||||||
|
$ tar -xf ../disk.img
|
||||||
|
$ ls
|
||||||
|
test.txt
|
||||||
|
```
|
||||||
|
|
||||||
|
## BSD tar / libarchive
|
||||||
|
|
||||||
|
Unfortunately, this broke bsdtar!
|
||||||
|
|
||||||
|
```
|
||||||
|
$ bsdtar -tf disk.img
|
||||||
|
bsdtar: Damaged tar archive
|
||||||
|
bsdtar: Error exit delayed from previous errors.
|
||||||
|
```
|
||||||
|
|
||||||
|
This is annoying because we run FreeBSD on the host for [opam.robur.coop](https://opam.robur.coop), our instance of [opam-mirror][opam-mirror].
|
||||||
|
This Autumn we updated [opam-mirror][opam-mirror] to use the hybrid GPT+tar GPTar *tartition table*[^tartition] instead of hard coded or boot parameter specified disk offsets for the different partitions - which was extremely brittle!
|
||||||
|
So we were no longer able to inspect the contents of the tar partition from the host!
|
||||||
|
Unacceptable!
|
||||||
|
So I started to dig into libarchive where bsdtar comes from.
|
||||||
|
To my surprise, after building bsdtar from the git clone of the source code it ran perfectly fine!
|
||||||
|
|
||||||
|
```
|
||||||
|
$ ./bsdtar -tf ../gptar/disk.img
|
||||||
|
test.txt
|
||||||
|
```
|
||||||
|
|
||||||
|
I eventually figure out [this change][libarchive-pr] fixed it for me.
|
||||||
|
I got in touch with Emily to let her know that bsdtar recently fixed this (ab)use of GNU volume headers.
|
||||||
|
Her reply was basically "as of when I wrote the article, I was pretty sure bsdtar ignored it."
|
||||||
|
And indeed it did.
|
||||||
|
Examining the diff further revealed that it ignored the GNU volume header - just not "correctly" when the GNU volume header was abused to carry file content as I did:
|
||||||
|
|
||||||
|
```diff
|
||||||
|
/*
|
||||||
|
* Interpret 'V' GNU tar volume header.
|
||||||
|
*/
|
||||||
|
static int
|
||||||
|
header_volume(struct archive_read *a, struct tar *tar,
|
||||||
|
struct archive_entry *entry, const void *h, size_t *unconsumed)
|
||||||
|
{
|
||||||
|
- (void)h;
|
||||||
|
+ const struct archive_entry_header_ustar *header;
|
||||||
|
+ int64_t size, to_consume;
|
||||||
|
+
|
||||||
|
+ (void)a; /* UNUSED */
|
||||||
|
+ (void)tar; /* UNUSED */
|
||||||
|
+ (void)entry; /* UNUSED */
|
||||||
|
|
||||||
|
- /* Just skip this and read the next header. */
|
||||||
|
- return (tar_read_header(a, tar, entry, unconsumed));
|
||||||
|
+ header = (const struct archive_entry_header_ustar *)h;
|
||||||
|
+ size = tar_atol(header->size, sizeof(header->size));
|
||||||
|
+ to_consume = ((size + 511) & ~511);
|
||||||
|
+ *unconsumed += to_consume;
|
||||||
|
+ return (ARCHIVE_OK);
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
So thanks to the above change we can expect a release of libarchive supporting further flavors of abuse of GNU volume headers!
|
||||||
|
🥳
|
||||||
|
|
||||||
|
[gptar-post]: gptar.html
|
||||||
|
[tar-executable]: https://uni.horse/executable-tarballs.html
|
||||||
|
[opam-mirror]: https://git.robur.coop/robur/opam-mirror/
|
||||||
|
[libarchive-pr]: https://github.com/libarchive/libarchive/pull/2127
|
||||||
|
|
||||||
|
[^tartition]: Emily came up with the much better term "tartition table" than what I had come up with - "GPTar".
|
55
src/blog.ml
55
src/blog.ml
|
@ -211,16 +211,18 @@ module Page = struct
|
||||||
method charset : string option
|
method charset : string option
|
||||||
method description : string option
|
method description : string option
|
||||||
method tags : string list
|
method tags : string list
|
||||||
|
method head_extra : string option
|
||||||
method with_host : string -> 'self
|
method with_host : string -> 'self
|
||||||
method get_host : string option
|
method get_host : string option
|
||||||
end
|
end
|
||||||
|
|
||||||
class page ?title ?description ?charset ?(tags = []) () =
|
class page ?title ?description ?charset ?(tags = []) ?head_extra () =
|
||||||
object (_ : #t)
|
object (_ : #t)
|
||||||
method title = title
|
method title = title
|
||||||
method charset = charset
|
method charset = charset
|
||||||
method description = description
|
method description = description
|
||||||
method tags = tags
|
method tags = tags
|
||||||
|
method head_extra = head_extra
|
||||||
val host = None
|
val host = None
|
||||||
method with_host v = {< host = Some v >}
|
method with_host v = {< host = Some v >}
|
||||||
method get_host = host
|
method get_host = host
|
||||||
|
@ -233,8 +235,9 @@ module Page = struct
|
||||||
let+ title = optional fields "title" string
|
let+ title = optional fields "title" string
|
||||||
and+ description = optional fields "description" string
|
and+ description = optional fields "description" string
|
||||||
and+ charset = optional fields "charset" string
|
and+ charset = optional fields "charset" string
|
||||||
and+ tags = optional_or fields ~default:[] "tags" (list_of string) in
|
and+ tags = optional_or fields ~default:[] "tags" (list_of string)
|
||||||
new page ?title ?description ?charset ~tags ()
|
and+ head_extra = optional fields "head_extra" string in
|
||||||
|
new page ?title ?description ?charset ~tags ?head_extra ()
|
||||||
|
|
||||||
let validate =
|
let validate =
|
||||||
let open Data.Validation in
|
let open Data.Validation in
|
||||||
|
@ -246,6 +249,7 @@ module Page = struct
|
||||||
("title", (option string) p#title);
|
("title", (option string) p#title);
|
||||||
("charset", (option string) p#charset);
|
("charset", (option string) p#charset);
|
||||||
("description", (option string) p#description);
|
("description", (option string) p#description);
|
||||||
|
("head_extra", option string p#head_extra);
|
||||||
]
|
]
|
||||||
end
|
end
|
||||||
|
|
||||||
|
@ -373,15 +377,17 @@ module Articles = struct
|
||||||
method title : string option
|
method title : string option
|
||||||
method description : string option
|
method description : string option
|
||||||
method articles : (Path.t * Article.t) list
|
method articles : (Path.t * Article.t) list
|
||||||
|
method head_extra : string option
|
||||||
method with_host : string -> 'self
|
method with_host : string -> 'self
|
||||||
method get_host : string option
|
method get_host : string option
|
||||||
end
|
end
|
||||||
|
|
||||||
class articles ?title ?description articles =
|
class articles ?title ?description ?head_extra articles =
|
||||||
object (_ : #t)
|
object (_ : #t)
|
||||||
method title = title
|
method title = title
|
||||||
method description = description
|
method description = description
|
||||||
method articles = articles
|
method articles = articles
|
||||||
|
method head_extra = head_extra
|
||||||
val host = None
|
val host = None
|
||||||
method with_host v = {< host = Some v >}
|
method with_host v = {< host = Some v >}
|
||||||
method get_host = host
|
method get_host = host
|
||||||
|
@ -418,11 +424,11 @@ module Articles = struct
|
||||||
|>> second
|
|>> second
|
||||||
(fetch (module P) ?increasing ~filter ~on ~where ~compute_link path)
|
(fetch (module P) ?increasing ~filter ~on ~where ~compute_link path)
|
||||||
>>> lift (fun (v, articles) ->
|
>>> lift (fun (v, articles) ->
|
||||||
new articles ?title:v#title ?description:v#description articles)
|
new articles ?title:v#title ?description:v#description ?head_extra:v#head_extra articles)
|
||||||
|
|
||||||
let normalize (ident, article) =
|
let normalize (ident, article) =
|
||||||
let open Data in
|
let open Data in
|
||||||
record (("url", string @@ Path.to_string ident) :: Article.normalize article)
|
record (("location", string @@ Path.to_string ident) :: Article.normalize article)
|
||||||
|
|
||||||
let normalize obj =
|
let normalize obj =
|
||||||
let open Data in
|
let open Data in
|
||||||
|
@ -431,32 +437,11 @@ module Articles = struct
|
||||||
; ("has_articles", bool @@ ((=) []) obj#articles)
|
; ("has_articles", bool @@ ((=) []) obj#articles)
|
||||||
; ("title", option string obj#title)
|
; ("title", option string obj#title)
|
||||||
; ("description", option string obj#description)
|
; ("description", option string obj#description)
|
||||||
|
; ("head_extra", option string obj#head_extra)
|
||||||
; ("host", option string obj#get_host)
|
; ("host", option string obj#get_host)
|
||||||
]
|
]
|
||||||
end
|
end
|
||||||
|
|
||||||
module Page_with_article = struct
|
|
||||||
class type t = object ('self)
|
|
||||||
inherit Page.t
|
|
||||||
method articles : (Path.t * Article.t) list
|
|
||||||
end
|
|
||||||
|
|
||||||
let normalize_article (ident, article) =
|
|
||||||
let open Data in
|
|
||||||
record (("url", string @@ Path.to_string ident) :: Article.normalize article)
|
|
||||||
|
|
||||||
let normalize (p : t) =
|
|
||||||
let open Data in
|
|
||||||
[
|
|
||||||
("title", (option string) p#title);
|
|
||||||
("charset", (option string) p#charset);
|
|
||||||
("description", (option string) p#description);
|
|
||||||
("tags", (list_of string) p#tags);
|
|
||||||
("articles", list_of normalize_article p#articles);
|
|
||||||
]
|
|
||||||
end
|
|
||||||
|
|
||||||
|
|
||||||
let is_markdown_file path =
|
let is_markdown_file path =
|
||||||
Path.has_extension "md" path ||
|
Path.has_extension "md" path ||
|
||||||
Path.has_extension "markdown" path
|
Path.has_extension "markdown" path
|
||||||
|
@ -518,7 +503,7 @@ struct
|
||||||
Pipeline.track_file Source.binary
|
Pipeline.track_file Source.binary
|
||||||
>>> Yocaml_yaml.Pipeline.read_file_with_metadata (module Article) file
|
>>> Yocaml_yaml.Pipeline.read_file_with_metadata (module Article) file
|
||||||
>>* (fun (obj, str) -> Eff.return (obj#with_host host, str))
|
>>* (fun (obj, str) -> Eff.return (obj#with_host host, str))
|
||||||
>>> Yocaml_cmarkit.content_to_html ()
|
>>> Yocaml_cmarkit.content_to_html ~strict:false ()
|
||||||
>>> Yocaml_jingoo.Pipeline.as_template
|
>>> Yocaml_jingoo.Pipeline.as_template
|
||||||
(module Article)
|
(module Article)
|
||||||
(Source.template "article.html")
|
(Source.template "article.html")
|
||||||
|
@ -541,7 +526,7 @@ struct
|
||||||
Articles.compute_index
|
Articles.compute_index
|
||||||
(module Yocaml_yaml)
|
(module Yocaml_yaml)
|
||||||
~where:is_markdown_file
|
~where:is_markdown_file
|
||||||
~compute_link:(Target.as_html @@ Path.abs [ "articles" ])
|
~compute_link:(Target.as_html @@ Path.abs [ "posts" ])
|
||||||
Source.articles
|
Source.articles
|
||||||
in
|
in
|
||||||
|
|
||||||
|
@ -556,10 +541,10 @@ struct
|
||||||
Yocaml_jingoo.render ~strict:true
|
Yocaml_jingoo.render ~strict:true
|
||||||
(List.map (fun (k, v) -> k, Yocaml_jingoo.from v) (Articles.normalize articles))
|
(List.map (fun (k, v) -> k, Yocaml_jingoo.from v) (Articles.normalize articles))
|
||||||
tpl))
|
tpl))
|
||||||
|
>>> Yocaml_cmarkit.content_to_html ~strict:false ()
|
||||||
>>> Yocaml_jingoo.Pipeline.as_template ~strict:true
|
>>> Yocaml_jingoo.Pipeline.as_template ~strict:true
|
||||||
(module Articles)
|
(module Articles)
|
||||||
(Source.template "layout.html")
|
(Source.template "layout.html")
|
||||||
>>> Yocaml_cmarkit.content_to_html ()
|
|
||||||
>>> drop_first ()
|
>>> drop_first ()
|
||||||
end
|
end
|
||||||
|
|
||||||
|
@ -570,7 +555,7 @@ struct
|
||||||
begin
|
begin
|
||||||
Pipeline.track_files [ Source.binary; file ]
|
Pipeline.track_files [ Source.binary; file ]
|
||||||
>>> Yocaml_yaml.Pipeline.read_file_with_metadata (module Page) file
|
>>> Yocaml_yaml.Pipeline.read_file_with_metadata (module Page) file
|
||||||
>>> Yocaml_cmarkit.content_to_html ()
|
>>> Yocaml_cmarkit.content_to_html ~strict:false ()
|
||||||
>>> Yocaml_jingoo.Pipeline.as_template ~strict:true
|
>>> Yocaml_jingoo.Pipeline.as_template ~strict:true
|
||||||
(module Page)
|
(module Page)
|
||||||
(Source.template "layout.html")
|
(Source.template "layout.html")
|
||||||
|
@ -594,7 +579,7 @@ struct
|
||||||
~filter
|
~filter
|
||||||
(module Yocaml_yaml)
|
(module Yocaml_yaml)
|
||||||
~where:is_markdown_file
|
~where:is_markdown_file
|
||||||
~compute_link:(Target.as_html @@ Path.abs [ "articles" ])
|
~compute_link:(Target.as_html @@ Path.abs [ "posts" ])
|
||||||
Source.articles
|
Source.articles
|
||||||
in
|
in
|
||||||
|
|
||||||
|
@ -609,10 +594,10 @@ struct
|
||||||
Yocaml_jingoo.render ~strict:true
|
Yocaml_jingoo.render ~strict:true
|
||||||
(List.map (fun (k, v) -> k, Yocaml_jingoo.from v) (Articles.normalize articles))
|
(List.map (fun (k, v) -> k, Yocaml_jingoo.from v) (Articles.normalize articles))
|
||||||
tpl))
|
tpl))
|
||||||
|
>>> Yocaml_cmarkit.content_to_html ~strict:false ()
|
||||||
>>> Yocaml_jingoo.Pipeline.as_template ~strict:true
|
>>> Yocaml_jingoo.Pipeline.as_template ~strict:true
|
||||||
(module Articles)
|
(module Articles)
|
||||||
(Source.template "layout.html")
|
(Source.template "layout.html")
|
||||||
>>> Yocaml_cmarkit.content_to_html ()
|
|
||||||
>>> drop_first ()
|
>>> drop_first ()
|
||||||
end
|
end
|
||||||
|
|
||||||
|
@ -626,7 +611,7 @@ struct
|
||||||
>>> Articles.fetch
|
>>> Articles.fetch
|
||||||
(module Yocaml_yaml)
|
(module Yocaml_yaml)
|
||||||
~where:(Path.has_extension "md")
|
~where:(Path.has_extension "md")
|
||||||
~compute_link:(Target.as_html @@ Path.abs [ "articles" ])
|
~compute_link:(Target.as_html @@ Path.abs [ "posts" ])
|
||||||
Source.articles
|
Source.articles
|
||||||
|
|
||||||
let rss1 =
|
let rss1 =
|
||||||
|
|
Loading…
Reference in a new issue