For those that missed it, last year I published a dive into the files and formats of the Titan A.E. demo. There's this and more available on our blog for those interested in #gamedev and #digipres, with more on the way!
https://talonbrave.info/2024/06/14/poking-titanae.html
niche question time! i'm trying to figure out what PREMIS event type would express "destroyed original carrier." If you use PREMIS and record this info, what value do you use? https://id.loc.gov/vocabulary/preservation/eventType.html #preservation #digipres
AI-bot scraper hinderung web archiving / web preservation
let me know!
48 vs 96 kHz for audio preservation?
I would still opt for 48kHz/24bit as perfectly sufficient.
In case of noisy/low-fi sources (Shellack, wax-cylinders, etc) one may choose 96kHz to have headroom for restoration.
Other than that, to capture regular-quality audio: Anyone got a good reason to go >48kHz?
Mini #digipres thread about invalid #PDF files, #JHOVE and PDF-HUL-114
( @BertrandCaron is leaving the old house, it's time to step up!)
At the beginning, there was a PDF 1.7 file donated to the #BnF and declared invalid by JHOVE 1.32.0. I start my investigation to find out if this is serious and if I'm going to repair the file (this is the way @mickylindlar told us to go)
1/...
The Xiph.Org Foundation has for many years hosted a collection of #lossless audio and video test clips to support the compression research community.
Unfortunately, that server may be going away soon. I'm looking at various options to keep the data online, but I wondered if it would be appropriate to upload a copy to @internetarchive The current collection is around 20 TB.
If so, what would be a good format? For clips originating in yuv, ffv1 or lossless vp9 compression is probably a good choice. (Uncompressed is usually best for benchmarking, but people can extract on their own.) Downloading multi-GB files is easier than it used to be. However, the highest quality for some clips are directories of png, tiff, or exr files. Not sure what to do with those. Tar them up? Will archive.org fall over if I upload 20,000 files as part of the same object?
Our March newsletter is out! We're pleased to share news of new #DigitalCulturalHeritage collections in the DRI repository, #OpenResearch developments in Ireland, and upcoming #DigiPres workshops at DRI.
Read the latest issue and subscribe https://dri.ie/dri-friends-newsletter/ #DigiPresNews
One of the car stereos I use has the strangest bug: Any #m4a file over 4:30 (4 minutes, 30 seconds) will skip for a second at 4:30, play for another 30 seconds or so, then just go onto the next track. Music from #iTunes or things from YouTube make no difference. Simply copying from .m4a to .aac with #ffmpeg avoids this issue. #digipres #aac
Occupé que j'étais à préparer ma transition vers la @tibhannover , je n'ai pas eu l'occasion de rappeler que l'échéance pour proposer un article #digipres à #iPRES2025 était le 14 avril.
Venez, on a des cookies, des autocollants à coller sur son ordinateur et surtout on est super sympas.
I truly believe it's time to take "thinking data differently" more serious and read up on Wikidata-related data-modeling stuff:
#Wikidata and the sum of all video games − 2024 edition: status update on our endeavour to become the hub of all video game metadata: 110,000 items, 70 new identifier properties, and a lot of video game genres.
https://commonists.wordpress.com/2025/03/24/wikidata-and-the-sum-of-all-video-games-2024-edition/
Hi fedi, another question re #SafeguardingResearch:
We got a couple of #Tableau data/visualizations that we need to archive.
But: the download is limited to Image, PDF or PowerPoint.
https://public.tableau.com/app/profile/dutytoserve/vizzes
Update: Currently looking into automating the step 'downloading PDF, with all sheets of the workbook'.
Ideas on how to accomplish this?
@lavaeolus For me personally, preserving a copy of the underlying dataset is priority number 1, because at least you have half a chance of reconstructing something from that. Then the source code of the site, which can either be used directly or at least give insight into the logic. The latter is definitely one for @swheritage
Neither of those captures the lived experience of using a web application though, which can be important in some instances.
Question re #SafeguardingResearch
We encounter 'web applications' that our current method of archiving don't preserve.
Things like [we need a better example, this one is already gone (but the data preserved) https://social.coop/@edsu/114206452552797815]
We are mostly using https://github.com/openzim/zimit to create WARC files and combining them into a single ZIM.
(This uses the browsertrix crawler)
Any ideas on how to archive not just the content, but also the functionality of such applications?
#DigiPres #Web #Archiving
I'm working on enabling the #DigiPres parts of our org by making some of the #COPTR Tool Grid apps available securely in a managed #Windows environment - first cab off the rank is a #Java app with bundled #JRE, which won't play well with our env
does anyone have experience using #maven & #jpackage as part of a build pipeline to create OS-native installers? my short-term target is a Windows MSI, but once that's in place the same pipeline should be able to spit out native installers for macOS (which we also use) & Linux (which we don't - yet
I'm at the point where I'm about to clone the repo & start tinkering
#boost4reach pls