Votre recherche
Résultats 19 ressources
-
Bibliothèque numérique de l’IRHT, répertoire des cotes de manuscrits, incunables et livres anciens conservés dans les bibliothèques du monde entier.
-
Overview This code in the R programming language downloads and processes the full set of resolutions, drafts and meeting records rendered by the United Nations Security Council (UNSC), as published by the UN Digital Library, into a rich and structured human- and machine-readable dataset. It is the basis for the Corpus of Resolutions: UN Security Council (CR-UNSC). All data sets created with this script will always be hosted permanently open access and freely available at Zenodo, the scientific repository of CERN. Each version is uniquely identified with a persistent Digitial Object Identifier (DOI), the Version DOI. The newest version of the data set will always available via the link of the Concept DOI: https://doi.org/10.5281/zenodo.7319780 Updates The CR-UNSC will be updated at least once per year. In case of serious errors an update will be provided at the earliest opportunity and a highlighted advisory issued on the Zenodo page of the current version. Minor errors will be documented in the GitHub issue tracker and fixed with the next scheduled release. The CR-UNSC is versioned according to the day of the last run of the data pipeline, in the ISO format YYYY-MM-DD. Its initial release version is 2024-05-03. Notifications regarding new and updated data sets will be published on my academic website at www.seanfobbe.com or on the Fediverse at @seanfobbe@fediscience.org Changelog New variant: EN_TXT_BEST containing a write-out of the English resolution texts equivalent to the CSV file text variable New diagrams: bar charts of top M49 regions and sub-regions of countries mentioned in resolution texts Fixed naming mix-up of BIBTEX and GRAPHML zip archives Fixed whitespace character detection in citation extraction (adds ca. 10% more citations) Fixed improper merging of weights in citation network Fixed "cannot xtfrm data frames" warning Improve REGEX detection for certain geographic entities Improve Codebook (headings, citation network docs) Functionality The pipeline will produce the following results and store them in the output/ folder: Codebook as PDF Compilation Report as PDF Quality Assurance Report as PDF ZIP archive containing the main data set as a CSV file ZIP archive containing only the metadata of the main data set as a CSV file ZIP archive containing citation data and metadata as a GraphML file ZIP archive containing bibliographic data as a BIBTEX file ZIP archive containing all resolution texts as TXT files (OCR and extracted) ZIP archive containing all resolution texts as PDF files (original and English OCR) ZIP archive containing all draft texts as PDF files (original) ZIP archive containing all meeting record texts as PDF files (original) ZIP archive containing the full Source Code ZIP archive containing all intermediate pipeline results ("targets") The integrity and veracity of each ZIP archive is documented with cryptographically secure hash signatures (SHA2-256 and SHA3-512). Hashes are stored in a separate CSV file created during the data set compilation process. System Requirements The reference data sets were compiled on a Debian host system. Running the Docker config on an SELinux system like Fedora will require modifications of the Docker Compose config file. 40 GB space on hard drive Multi-core CPU recommended. We used 8 cores/16 threads to compile the reference data sets. Standard config will use all cores on a system. This can be fine-tuned in the config file. Given these requirements the runtime of the pipeline is approximately 40 hours. Instructions Step 1: Prepare Folder Copy the full source code to an empty folder, for example by executing: $ git clone https://github.com/seanfobbe/cr-unsc Always use a dedicated and empty (!) folder for compiling the data set. The scripts will automatically delete all PDF, TXT and many other file types in its working directory to ensure a clean run. Step 2: Create Docker Image The Dockerfile contains automated instructions to create a full operation system with all necessary dependencies. To create the image from the Dockerfile, please execute: $ bash docker-build-image.sh Step 3: Compile Dataset If you have previously compiled the data set, whether successfuly or not, you can delete all output and temporary files by executing: $ Rscript delete_all_data.R You can compile the full data set by executing: $ bash docker-run-project.sh Results The data set and all associated files are now saved in your working directory. GNU General Public License Version 3 Copyright (C) 2024 Seán Fobbe, Lorenzo Gasbarri and Niccolò Ridi This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public Licensealong with this program. If not, see https://www.gnu.org/licenses/ Author Websites Personal Website of Seán Fobbe Personal Website of Lorenzo Gasbarri Personal Website of Niccolò Ridi Contact Did you discover any errors? Do you have suggestions on how to improve the data set? You can either post these to the Issue Tracker on GitHub or contact Seán Fobbe via https://seanfobbe.com/contact/
-
Diving into CSR's launch of a clear, concise repository of deployed nuclear weapons across their history.
-
Digitised historic editions of the Lloyd's Register of Ships, including the oldest surviving Register Book from 1764.
-
Bibliothèque numérique du Bureau international du Travail. Diffuse les publications de l’OIT (livres, articles, rapports, manuels) et recense toute autre documentation portant sur le monde du travail.
-
An Awesome List of computer history videos, documentaries and related folklore
-
The David Rumsey Map Collection was started over 35 years ago and contains more than 200,000 maps. The collection focuses on rare 16th through 21st century maps of North and South America, as well as maps of the World, Asia, Africa, Europe, and Oceania. The collection includes atlases, globes, wall maps, school geographies, pocket maps, books of exploration, maritime charts, and a variety of cartographic materials including pocket, wall, children's, and manuscript maps. Items range in date from around 1550 to the present.
-
Index of place records with coordinates, historical metadata, linked toponyms, and ecological classifications
-
Find books and articles from university and national libraries from all around the world with the OpenCat meta catalog.
-
OpenHistoricalMap est un projet conçu pour stocker et afficher des données cartographiques tout au long de l’histoire du monde. Il s’agit d’un travail en cours, nous jouerons avec de nombreuses nouvelles fonctionnalités au fur et à mesure que nous mettrons en place le site. Nous vous encourageons également à commencer à jouer autour et à modifier les données.
-
The easy-to-use getaway to historical maps in libraries around the world.
-
The most complete library of Marxism with content in over 80 languages and the works of over 1000 authors readily accessible by archive, subject, or history as well as hundreds of periodicals.
Explorer
Lieux
Sujets
- Droit et lois (1)
- Economie (1)
- Études noires (1)
- Militaire et paix (1)
- Politique (4)
- Sociologie & Travail (1)
Types
- Actualités (1)
- Cartes (5)
- Informatique (4)
- Livres (1)